• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    CrossFormer Embedding DeepLabv3+for Remote Sensing Images Semantic Segmentation

    2024-05-25 14:42:28QixiangTongZhipengZhuMinZhangKeruiCaoandHaihuaXing
    Computers Materials&Continua 2024年4期

    Qixiang Tong,Zhipeng Zhu,Min Zhang,Kerui Cao and Haihua Xing

    School of Information Science and Technology,Hainan Normal University,Haikou,571158,China

    ABSTRACT High-resolution remote sensing image segmentation is a challenging task.In urban remote sensing,the presence of occlusions and shadows often results in blurred or invisible object boundaries,thereby increasing the difficulty of segmentation.In this paper,an improved network with a cross-region self-attention mechanism for multi-scale features based on DeepLabv3+is designed to address the difficulties of small object segmentation and blurred target edge segmentation.First,we use CrossFormer as the backbone feature extraction network to achieve the interaction between large-and small-scale features,and establish self-attention associations between features at both large and small scales to capture global contextual feature information.Next,an improved atrous spatial pyramid pooling module is introduced to establish multi-scale feature maps with large-and small-scale feature associations,and attention vectors are added in the channel direction to enable adaptive adjustment of multi-scale channel features.The proposed network model is validated using the Potsdam and Vaihingen datasets.The experimental results show that,compared with existing techniques,the network model designed in this paper can extract and fuse multiscale information,more clearly extract edge information and small-scale information,and segment boundaries more smoothly.Experimental results on public datasets demonstrate the superiority of our method compared with several state-of-the-art networks.

    KEYWORDS Semantic segmentation;remote sensing;multiscale;self-attention

    1 Introduction

    With the rapid development of aerospace technology and integrated earth observation systems,we are entering the era of big remote sensing data.Remote sensing technology is widely used in ecological environment monitoring [1],precision agriculture [2,3],land and resource surveys [4],and urban planning [5,6].The explosion of remote sensing data makes the efficient processing and intelligent interpretation of massive remote sensing data a critical problem[7].Semantic segmentation of remote sensing images,which is an important topic of remote sensing image processing,is a prerequisite for subsequent scene understanding,feature monitoring,and 3D reconfiguration,and has played a significant role in promoting the development of remote sensing technology.

    In recent years,the powerful capabilities of convolutional neural networks have been extensively investigated.Because of their robust feature extraction capabilities,feature extraction modes from superficial to deep have enabled the conversion from shallow detailed features (e.g.,color,location,texture) to higher abstract categories,realizing the effective capture of semantic information [8,9].Therefore,the use of deep learning-based methods to segment feature information quickly and accurately has become a hot topic of research.In 2015,Long et al.proposed the Fully Convolutional Network(FCN)[10],which achieved end-to-end semantic segmentation for the first time and applied convolutional neural networks to the field of image semantic segmentation.Subsequently,SegNet[11]and DeconvNet [12] were proposed to reduce the loss of detail in FCNs.Ronneberger et al.further optimized FCNs in 2015 by proposing a U-Net [13] network model based on the encoder and decoder structures.U-Net was first applied to medical image semantic segmentation,which has an encoder similar to the FCN and extracts image features through operations such as concession and pooling.Yu et al.proposed the Dilated Convolution[14],which retains more contextual information by expanding the perceptual field while maintaining a constant resolution.The nature of the local receptive field of the full convolutional network restricts the pixel-level recognition to the local region.This makes it difficult to connect the contextual cues,and affects the further improvement of the segmentation accuracy.

    In addition to the above,Chen et al.proposed the DeepLab v3 [15] network model based on the encoder and decoder structure,which improved on the previous DeepLab v1 [16] and DeepLab v2[17]network models by obtaining multiscale context information with the help of several parallel dilated convolutions with different dilation rates.The segmentation accuracy of the network model is improved with the help of several parallel convolutions with different atrous rates to obtain multiscale context information.Following that,they further introduced DeepLabv3+[18].Zhao et al.proposed the Pyramid Scene Parsing Network(PSPNet)[19]with several global pooling operations of different step lengths to complete the aggregation of multiscale contextual information.Zhang et al.[20]achieved the fusion of high-and low-level features by constructing a feature pyramid network that propagates features from bottom to top.Chu et al.[21] enhanced the recognition accuracy of small targets by utilizing the fusion of features from multiple convolutional layers.Extracting multiscale information allows the model to better understand and process different scales and structures within images,thereby improving the performance of image analysis and processing tasks.This is particularly important in the fields of semantic segmentation,instance segmentation,and object detection.The methods of Zhang et al.[20]and Chu et al.[21]expanded the receptive field,and thus improved the coarse results of feature extraction.However,it is difficult to gain fine-grained global information from remote sensing images with complex backgrounds.

    In recent years,convolutional neural network-based approaches have dominated various tasks in computer vision.With the emergence of the Transformer [22] model,the self-attention mechanism has been widely used and pushed to a higher level.Initially,Transformer was introduced to the field of natural language processing.Inspired by the successful application of Transformer to natural language processing,many researchers tried to apply Transformer to computer vision tasks.In 2020,Dosovitskiy et al.proposed the Vision Transformer(ViT)[23]model,which achieved excellent results on image classification tasks.The ViT model usually requires a huge amount of data for pre-training and migrates to small tasks for classification recognition,which is more scalable than traditional convolutional neural networks.In 2021,Han et al.proposed the Transformer in Transformer(TNT)[24]model,which improves on the ViT model by better extracting the global features and local features of the image.Xie et al.[25] proposed SegFormer,which achieves efficient and high-performance segmentation.Touvron et al.proposed the DeiT (Data-efficient image Transformers) [26] model,which extends ViT in terms of knowledge distillation.Wang et al.improved Transformer in 2021 and proposed the CrossFormer [27] model based on the cross-region self-attention mechanism,which can better capture global contextual feature information.Zhao et al.[28] and Li et al.[29]presented region-level attention and frame-level attention for video captioning.Zhao et al.[30]explored the effectiveness of pairwise self-attention and patchwise self-attention in image recognition.SENet [31] expresses the relationship between channels through the global average pooling layer to understand the importance of different channels automatically.Cao et al.[32]employed a hierarchical SwinTransformer architecture and introduced Swin-Unet for medical image segmentation.These new models have made great progress in natural language processing and medical image segmentation,but their segmentation potential when applied to remote sensing images requires further analysis.Different from the above methods,we consider the self-attention associations between features at both large and small scales to capture global contextual feature information.

    High-resolution remote sensing images (RSIs) contain rich feature information and complex backgrounds.The complete semantic information understanding is the basis of high-resolution RSI segmentation.Existing segmentation methods for high-resolution RSIs suffer from problems such as blurred segmentation edges and difficulty in segmenting small-scale features [4].In particular,the presence of occlusions and shadows often leads to blurred or invisible object boundaries,which increases the difficulty of segmentation [33,34].By learning information at different scales of the image,the complementary information capability of the network can be enhanced,and the loss of feature details and edge information can be avoided to a certain extent.Therefore,this paper describes a network model with multiscale features and the cross-region self-attention mechanism,and combines this with the advantages of the Transformer,which can solve the problems of smallobject segmentation and blurred target edge segmentation in the semantic segmentation of RSIs.First,the cross-regional self-attention mechanism is incorporated into the encoder of the network,and selfattention is established to determine the correlation between features at different scales,thus capturing global contextual feature information.A modified atrous spatial pyramid pooling(ASPP)module is then introduced after the cross-regional self-attention mechanism to increase the image perceptual field and capture multiscale image feature information.The specific network architecture and design ideas are described in detail in the following sections.

    The main contributions of this paper are as follows:

    (1) We design a network model with multiscale features and the cross-region self-attention mechanism to address the challenges of segmenting small objects.This model improves the accuracy of object edge segmentation in high-resolution urban remote sensing imagery.

    (2) We incorporate cross-scale embedding layers and cross-region self-attention mechanisms into DeepLabv3+,linking features of different scales with contextual information to enhance spatial attention.

    (3) We improve the ASPP module to enhance the model’s ability to extract edge information and strengthen multiscale representation.The proposed network employs five parallel convolution kernels with different receptive field sizes to extract features.We optimize the network parameters in ASPP using spatially separable convolutions in place of traditional convolutions.

    2 Related Work

    In this section,we review existing theories and methods related to our proposed model.This includes the self-attention mechanism and the CrossFormer.

    2.1 Semantic Segmentation of Remote Sensing Images

    Semantic segmentation of RSIs continues to receive widespread attention.Most current segmentation methods focus on the effective fusion of low-level spatial details and high-level semantic cues,or incorporate boundary supervision to obtain boundary guidance.Remote sensing is enriched with spatial information,but this may be masked when high-level features and lower feature layers are fused.To address this problem,Wen et al.[35]proposed the MLWNet semantic segmentation model,which uses multiscale linear self-attention modules to abstract the correlation between contexts.To enhance the contextual information and alleviate the semantic ambiguity resulting from occlusion,Xiao et al.[36] used a parallel context aggregation module in the encoder and a feature shrinkage module in the downsampling process to minimize the information loss.Wang et al.[37]investigated building extraction using high-resolution RSIs from three aspects: The extraction of multiscale information,the multilevel merging of high-level semantic information,and multilevel information fusion.To avoid any loss of spatial detail and improve the segmentation accuracy,Fan et al.[38]proposed a progressive adjacent-layer coordination symmetric cascade network for the cross-layer fusion of multimodal remote sensing data and the preservation of spatial information.For effective aggregation of remote contextual cues and combining multilevel features,Wang et al.[39]developed a boundary guided multilevel feature fusion module.This module embeds the boundary guided information into the multilevel feature fusion process and subtly facilitates spatial contextual and channel-level semantic correlation at the pixel level.Wang et al.[40] constructed an asymmetrical convolution and orientation attention module for the adaptive selection of favorable segmentation features and enhancement of the inherent geometric features of remote sensing target contours.Using the multiscale features of the image effectively improves the coarse results of feature extraction and fuses the global contextual semantic features,but tends to lose some boundary information.Therefore,the extraction of cross-scale information of RSIs,the recovery of object boundary information,and the localization of small-scale features require further study.In particular,effective auxiliary modules for deep learning models need to be developed.

    2.2 Self-Attention Mechanism

    Self-attention can be considered as a network layer that can better correlate contextual information in the input vectors than fully connected neural networks.In general,the network training needs to consider the entire set of input vectors.Fully connected neural networks require the connection of all input vectors,which generates a large number of parameters and leads to overfitting,as well as increasing the computational effort.Unlike fully connected neural networks,self-attention has a larger field of perception and achieves the association between each pixel in an image through a vector group.

    There is a one-to-one correspondence between the self-attention input vector and the output vector,as shown in Fig.1a.The input sequence isXand the output sequence isY.FC is the fully connected layer.The self-attention mechanism combines the input vectorXwith contextual information and then outputsYwith global relevance through FC.The specific process can be described by the following equation:

    where ? is the fully connected layer FC and ω is the self-attention mechanism layer.

    Figure 1: (a) Schematic diagram of the self-attention mechanism.(b) Operation flow of scaled dotproduct for self-attention mechanism

    The calculation process of the self-attention mechanism is shown in Fig.1b.In the proposed method,a scaled dot-product self-attention mechanism is used.The dot-products of the input sequenceXwith the matricesUq,Uk,Uvgive the matricesQ,K,andV,respectively.The specific calculation process is expressed by the following equations:

    where Uq,Uk,and Uvare weight matrices that are updated by network training.Qis called the Query andK-Vis the Key-Value pair.Ais called the attention score,which indicates the similarity betweenQandK.SoftMax is the fully connected layer.Yindicates the weighted aggregation byAandV.dkdenotes the dimension ofK.

    2.3 Cross-Regional Self-Attention Mechanism

    Transformer is a deep neural learning network based on a self-attention mechanism that processes data in parallel.It slices the input image into patches of equal size,spreads the patches into onedimensional vectors by embedding,and labels each patch with a positional encoding.This ensures that all embeddings in the same layer of the network model have the same scale and removes cross-scale characterization.Based on the powerful global modeling capability of Transformer,Wang et al.[27]developed CrossFormer based on the cross-region self-attention mechanism,establishing the Crossscale Embedding Layer(CEL)and the Long Short Distance Attention(LSDA)module.The LSDA module can better capture global contextual feature information.

    Fig.2a shows the overall CrossFormer structure and Fig.2b shows two consecutive CrossFormer modules.Each CrossFormer module consists of a Short-Distance Attention (SDA) module or a Long-Distance Attention (LDA) module and a multi-layer perceptron (MLP).Moreover,the SDA and LDA modules alternate in consecutive CrossFormer modules,while the Dynamic Position Bias(DPB) module acts on the SDA and LDA to obtain the embedded position representation.Layer Normalization (LN) is applied before each module and residual connectivity is applied after each module.

    Figure 2: (a) Structure of crossformer for image segmentation.(b) Two consecutive crossformer blocks.SDA and LDA represent long distance attention mechanisms and short distance attention mechanisms,respectively.DPB represents dynamic position bias

    2.3.1 Cross-Scale Embedding Layer

    The cross-scale embedding layer fuses multiple image blocks of different scales to provide cross-scale embedding for each stage of the input to the network model.As shown in Fig.3,four convolutional kernels of different sizes and equal steps are set to sample the feature maps,resulting in four feature maps of different sizes.These four feature maps are projected and connected to generate the embedding layer on the right.Each embedding contains feature information of four different scales.For cross-scale embedding,setting the projection dimension of each scale is a key issue.The computational effort of a convolutional layer is proportional toK2D2,in whichKis the size of the convolutional kernel andDis the dimension of the input or output(assuming that the input dimension is equal to the output dimension).A large convolutional kernel implies more computational effort than a small one for a given dimensionality.Therefore,the embedding layer employs different dimensional stitching for convolutional kernels of different sizes to balance the computational cost with the model performance.

    Figure 3: Cross-scale embedding layer of CrossFormer

    The cross-scale embedding layer can be realized by four convolutional layers,as described in Table 1.Four convolutional kernels of different sizes(4×4,8×8,16×16,and 32×32)are adopted for the embedding layer.Assuming that the embedding has a total of 256 dimensions,it can be found from the table that larger convolutional kernels use fewer dimensions,while smaller ones use more dimensions.

    Table 1: Parameters of the cross-scale embedding layer

    2.3.2 LSDA Mechanism

    Self-attention involves computing the relationship between each patch token and all other tokens.This intensive computation leads to an increase in the overhead of the model.To reduce the computational load,the self-attention module is divided into SDA and LDA units.

    SDA divides the adjacent image blocks into one group,as shown in Fig.4a.All image blocks in the red border are divided into one group,which uses a window of size(in Fig.4a,n=3).Each group has nine image blocks,so there are a total of nine groups.SDA restricts the self-attention calculation to within each group,thus obtaining the relevance of the local information.LDA performs interval sampling,as shown in Fig.4b,using a fixed interval of i=3 for sampling.Therefore,the image blocks represented by red,blue,and green borders are divided into separate groups.LDA acquires cross-regional information because its self-attention computation is cross-scale,unlike SDA.LDA reduces the computational volume and retains feature information from both small and large scales.The calculation process of DPB is shown in Fig.4c.

    Figure 4: (a)Short Distance Attention(SDA).Embeddings(yellow squares)are grouped by red boxes.(b) Long Distance Attention (LDA).Embeddings with the same borders belong to the same group.(c)Dynamic Position Bias(DPB),with its output being a scalar

    3 Proposed Method

    DeepLab was launched in 2014 and has shown strong potential in semantic segmentation.The rise of ViT has taken visual recognition into a new era.Hierarchical Transformer structures(e.g.,CrossFormer,SwinTransformer [41]) have enabled the success of ViT as a generalized visual task backbone,and have shown strong performance in major visual tasks.The powerful global modeling ability of Transformer is what CNNs lack.Therefore,we combine CrossFormer with a CNN.CrossFormer solves the problem of insufficient interaction between different scale features,and achieves the correlation of information at different scales,establishing contextual connections.In revisiting the development of computer vision,ConvNeXt[42]shows us that ConvNets are highly dynamic.We believe that DeepLab still has research value and have explored it in conjunction with Transformer.

    3.1 Multiscale Feature Cross-Regional Self-Attention Mechanism Network

    The proposed network with multiscale features and the cross-region self-attention mechanism is shown in Fig.5.The overall structure follows the design of DeepLabv3+.It inherits the encoding–decoding structure of DeepLabv3+and realizes the fusion of shallow features and multiscale features.

    Unlike ViT,CrossFormer compensates for the interaction of different-scale features.The extraction of equal-scale features in each layer and the adjacent embedding of self-attention modules not only uses cross-scale interaction,but also sacrifices small-scale features.The core of CrossFormer,i.e.,the Cross-scale Embedding Layer(CEL),and LSDA generate a cross-attention module that realizes the associations of different-scale information in Transformer.This is in accordance with the idea of the DeepLabv3+backbone using atrous convolution to extract different-scale features.Thus,we consider the CrossFormer Block as the core for feature extraction.

    The backbone structure used in the proposed method is shown in Fig.6.First,the backbone network extracts features from the input map,which is a cross-region self-attention network consisting of four CrossFormer-Stages connected sequentially.Each CrossFormer-Stage consists of a CEL and a CrossFormer Block(see Fig.2a).The CEL mixes each embedding with multiple patches of different scales (see Fig.3).In the backbone,the feature map (Feature-a) output from the first stage is used as a low-level feature input to the decoder for skip connection.In the cross-regional self-attention mechanism network,the feature map output from the first stage is input into the decoder as a low-level feature for hopping connections.Feature-a implements feature fusion with the high-level semantics of another branch in the decoder.Feature-b,which is output from the last stage of the backbone network,is input into the depthwise-separable ASPP as high-level semantic features.We use bilinear interpolation upsampling throughout the network.

    Figure 5: Modified model based on DeepLabv3+designed in this paper

    Figure 6: CrossFormer backbone network described in this paper

    3.2 Depthwise-Separable ASPP Module with CBAM

    SeNet,ECA (Efficient Channel Attention) [43],and SE Block (Squeeze-and-excitation) [44]construct interdependencies between channels,allowing the network to adaptively recalibrate feature responses in the channel direction.ConvNets extract feature information by mixing cross-channel and spatial information,and the Convolutional Block Attention Module (CBAM) [45] emphasizes the features along these two dimensions:Channels and spatial axes.CBAM applies the channel and spatial attention modules sequentially(see Fig.7),enabling each branch to learn“what”and“where”in the channels and spatial axes,respectively.The attention is inferred along these two dimensions to refine the image features adaptively.Thereby,the network can concentrate on essential features and suppress the unwanted features.

    Figure 7: (a) Overall structure of CBAM.(b) Channel attention mechanism module in CBAM.(c)Spatial attention mechanism module in CBAM

    We introduce CBAM to further refine and supervise the multiscale feature layer.The spatial attention module can supervise the spatial features of the extracted multiscale feature layers under different expansion rates,which enables the model to focus its attention on specific regions of the image,thus improving the model perception and adapting to inputs of different scales.The channel attention module enables the model to concentrate on channels containing specific features of greater importance,and dynamically adjusts the attention to each channel,thus reducing the influence of redundant channels and improving the expression ability of features.Therefore,we use CBAM after ASPP to supervise the multiscale feature map in two dimensions and further enhance the connection between large and small scales.

    We apply Depthwise-Separable Convolution (DSC) to the ASPP module to constitute a depthseparable ASPP.DSC includes both Depthwise Convolution and Pointwise Convolution,as shown in Fig.8.

    Figure 8: Schematic diagram of Depthwise-Separable Convolution(DSC)

    The input image or feature map is subjected to Depthwise Convolution to obtain new feature maps with the same number of channels as the input layer.To add the number of feature maps,a new feature map is generated by splicing the feature map dimensionally by Pointwise Convolution(1×1 Conv).DSC integrates both the spatial information and the information from different channels;compared with standard convolution,it has fewer parameters and lower computational cost.

    Compared with the original DeepLabv3+backbone network,the CrossFormer backbone has cross-scale representation,but does not possess the strengths of atrous convolution for edge information extraction.Thus,we use five different 3 × 3 DSCs (with rates of 1,6,12,18,and 24) in the depthwise-separable ASPP module to enhance the multiscale representation.ASPP can encode multiscale contextual information with filtering or pooling.The loss of boundary information during the downsampling of the backbone network can be alleviated by extracting denser feature mappings with atrous convolution.The ASPP module designed for the experiments in this paper is shown in Fig.9.

    Figure 9: Diagram of depthwise-separable ASPP module

    The main improvements of this ASPP module are described below.First,we add an atrous convolution with a rate of 24 to the original ASPP module.The purpose of this improvement is to obtain a larger receptive field and provide a more effective extraction of the fine-edge information that is easily lost in the downsampling process.Second,to improve the efficiency of model training,depthwise-separable atrous convolution is introduced by replacing the convolution in the original ASPP module.The new ASPP module has a 1×1 DSC in the first branch,which is designed to preserve the original receptive field.The second to fifth branches apply depthwise-separable atrous convolution with different rates to obtain different receptive fields for multiscale feature extraction.Third,global average pooling is used to obtain global feature information.Finally,the feature maps are stacked in the channel dimension and passed through a standard(1×1)convolution to fuse the information of different scales.The final deep information feature map of the encoder is output through the CBAM.

    4 Experiments and Results Analysis

    4.1 Datasets

    The experiments are conducted using two open datasets,Potsdam and Vaihingen,provided by the International Society for Photogrammetry and Remote Sensing(ISPRS)[46].Potsdam is a highresolution RSI dataset containing 38 aerial high-resolution RSIs taken by an unmanned aerial vehicle over the German city of Potsdam.Each image is 6000×6000 pixels in size.Vaihingen is an aerial image set of a German village,containing 33 aerial images of different sizes.The numbers of images used for training,validation,and testing are listed in Table 2.Sample images of the Potsdam and Vaihingen datasets are shown in Fig.10.

    Table 2: Division of the Potsdam and Vaihingen datasets

    Both the Potsdam and Vaihingen datasets were classified into six common landcover categories,namely Impervious surfaces,Buildings,Low vegetation,Trees,Cars,and Background,as described in Table 3.

    Table 3: Potsdam and Vaihingen datasets category label color comparison table

    Figure 10: (a) Potsdam dataset image with corresponding labels.(b) Vaihingen dataset image with corresponding labels

    4.2 Data Preprocessing

    In this experiment,the original large images of the Potsdam and Vaihingen datasets and the corresponding segmentation labels were cropped into sub-images of size 512 × 512 pixels.The cropping selection was performed in sliding window mode,with the image edge repetition rate set to 15%,as shown in Fig.11.

    To further enhance the model’s generalization ability,we expanded the number of samples using data augmentation.This increases the number of samples by adding Gaussian noise or pepper noise,or by performing random rotation,vertical flip,and horizontal flip operations(Fig.12).Data augmentation improves the generalization ability of the model.

    4.3 Evaluation Indicators

    To objectively evaluate the segmentation effect of each model,this paper uses the Confusion Matrix (Table 4) to calculate the Intersection over Union (IoU),Mean Intersection over Union(MIoU),and F1-score of the segmentation results given by each model.

    Table 4: Confusion matrix

    Figure 11: (a)and(b)represent the cropped RSI and their labels,respectively

    Figure 12: Data enhancement effect

    The evaluation indicators are defined as follows:

    where TP,FP,FN,and TN represent True Positive,False Positive,False Negative,and True Negative,respectively.Precision is the percentage of positive samples that are correctly predicted in the prediction results.Recall is the percentage of predicted true samples out of the total number of true samples.Accuracy is the ratio of correct predictions to the total number of predicted samples,and the F1-score is an overall index combining the Precision and Recall.IoU is the ratio of the intersection of predicted and true values of a category to the concurrent set,and MIoU is the average of the ratio of the intersection and concurrent set of all categories.

    In addition,to evaluate the time and space complexity of the network,the number of parameters in each model and the number of floating point operations(FLOPs)are compared.

    4.4 Experimental Environment and Parameters

    The software and hardware configurations used for the experiments are listed in Table 5.

    Table 5: Experimental hardware and software configuration table

    The optimizer uses the stochastic gradient descent(SGD)algorithm with a“Poly”learning rate strategy,and sets the initial learning rate to 0.01,momentum to 0.9,and weight decay to 0.0005.The training Batch Size is set to 8,with a total of 80,000 iterations.The loss value of the network model and the accuracy of each category are output every 40 iterations.We used cross-entropy loss functions in our experiments.

    4.5 Analysis of Experimental Results

    To evaluate the performance of the network models designed in this paper,we cropped the large images from the Potsdam and Vaihingen datasets into training images of 512 × 512 pixels for the experiments,and quantitatively evaluated and compared U-Net,DeepLabv3+,SwinUnet,SegFormer,TCUnet[47]and DBENet[48]against the models proposed in this paper.

    From Table 6,it can be found that the MIoU of the proposed model reached 76.41% on the Potsdam dataset,and the mean F1-score (mFscore) reached 86.73%.Compared with U-Net,DeepLabv3+,SwinUnet,SegFormer,TCUnet and DBENet,this represents an improvement of 8.74%,6.95%,9.40%,0.98%,5.65%and 3.63%in MIoU and 6.96%,5.30%,8.35%,1.36%,6.04%and 5.41%in mFscore.Compared with the IoU of DeepLabv3+,the proposed network model gives an improvement of 7.29%,5.45%,9.45%,11.18%,6.74%,and 1.59% in the six landcover categories of impervious surfaces,buildings,low vegetation,trees,cars,and background,respectively.The specific quantitative evaluation indicators are listed in Tables 6 and 7.

    Table 6: Comparison of overall evaluation metrics of different network models on the Potsdam dataset(%)

    Table 7: IoU(%)for different network models on the Potsdam datasets

    From Table 8,it can be found that the MIoU of the proposed network model reached 88.95%on the Vaihingen dataset,and the mFscore reached 94.08%.Compared with U-Net,DeepLabv3+,SwinUnet,SegFormer,TCUnet and DBENet,this represents an improvement of 7.86%,3.79%,0.72%,0.83%,2.36% and 1.93% in MIoU and 4.72%,2.29%,0.52%,0.50%,1.87% and 1.92% in mFscore.Compared with the IoU of DeepLabv3+the proposed network model produces an improvement of 1.98%,1.66%,1.44%,0.93%,7.45%,and 6.26%in the six landcover categories of impervious surfaces,buildings,low vegetation,trees,cars,and background,respectively.The specific quantitative evaluation indicators are presented in Tables 8 and 9.

    Table 8: Comparison of overall evaluation metrics for different network models on the Vaihingen datasets(%)

    Table 9: IoU(%)for different network models on the Vaihingen datasets

    From Tables 6 and 8,it can be found that the experimental results obtained for the Vaihingen dataset are similar to those obtained for the Potsdam dataset.This proves that the proposed network model is effective in introducing a cross-regional self-attention mechanism and an improved depthwise-separable ASPP to extract cross-scale features and enhance the performance of the network model.

    The segmentation results of the various models on the Potsdam and Vaihingen datasets are shown in Figs.13 and 14.Looking at Fig.13,we see that the models (such as Unet,DeepLabv3+,and SwinUnet) have relatively vague segmentation boundaries for buildings (labeled with blue in Fig.13).For the segmentation of low vegetation(cyan)and trees(green),the models(such as Unet,DeepLabv3+,TCUnet,and DBENet) can achieve the segmentation of the main parts,but the boundary between the two categories is still unclear.The proposed model exhibits greater sensitivity in identifying low vegetation and trees.The segmentation boundaries are more refined than in the other six methods,and the position localization is more accurate.We see that the proposed model provides better segmentation than the other six models,especially in the red rectangular boxes.From Fig.14,it can be observed that the other six models produce a certain degree of fuzziness at the edges of buildings and smaller-scale low vegetation.The proposed model effectively alleviates this issue and accurately locates small-scale information.The experimental results show that the model described in this paper is more detailed in the segmentation of high-resolution RSI datasets,captures small-scale information more accurately,and reproduces the details better.

    4.6 Ablation Experiment

    We denote the CrossFormer backbone network,the ASPP module,and the CBAM as modules 1,2,and 3,respectively.To verify the performance of each component of the whole network,we conducted ablation experiments on the Potsdam dataset.

    “Deeplabv3+” represents the original DeepLabv3+model.“Deeplabv3++1+2” denotes the original backbone network and ASPP in DeepLabv3+have been replaced by modules 1 and 2.We conducted a longitudinal comparison to analyze the performance of the proposed modules.

    Comparison between“Deeplabv3++all”and“Deeplabv3++1+3”(where the former utilizes an improved ASPP module and the latter uses the original ASPP module):After incorporating modules 1 and 3,the improved ASPP module exhibits an improvement of 1.73%in MIoU and 1.71%in mFscore compared to the original ASPP module.

    Figure 13: Visualization results of Potsdam datasets

    Figure 14: Visualization results of Vaihingen datasets

    “Deeplabv3++all”compared to“Deeplabv3++2+3”(where the former uses the CrossFormer backbone network and the latter employs the original convolutional network): After replacing the backbone network,there was a noticeable improvement.Based on the metrics,the CrossFormer-based structure shows an increase of 3.85% in MloU and 3.79% in mFscore compared to the CNN-based structure.

    “Deeplabv3++all”compared to“Deeplabv3++1+2”(where the former incorporates CBAM,while the latter does not use CBAM):CBAM focuses on crucial feature information,enhancing the model’s perception of multiscale features.In terms of MIoU and mFscore,“Deeplabv3++all”gives improvements of 1.6%and 1.77%,respectively,over“Deeplabv3++1+2”.

    The results of the ablation experiments are shown in Fig.15.From the data in Table 10,the combination of modules 1+2 and 1+3 significantly improves the accuracy.In Fig.15,the 1 +2 combination significantly improves the extraction of cars and low vegetation,and the extracted boundaries are smooth and clear.This indicates that CrossFormer and the improved ASPP enhance the ability to recognize small-scale features.The combination of modules 1+3 is highly effective in recognizing buildings with fewer“voids”.This indicates that our enhancement of cross-scale feature information exchange is effective.

    Table 10: Ablation experimental design

    Figure 15: Examples of semantic segmentation results on the Potsdam dataset.a.Deeplabv3++1+2.b.Deeplabv3++2+3.c.Deeplabv3++1+3

    5 Conclusions

    This paper has described a cross-regional multiscale DeepLabv3+improvement model for the difficult problem of small object segmentation in RSIs.The cross-scale embedding layer was established by replacing the backbone feature extraction network of DeepLabv3+with CrossFormer,and a crossregional self-attention mechanism was introduced to enhance the connection between local and global contextual information.A modified ASPP structure with CBAM was then used to extract multiscale features to enhance the model’s ability to recognize small objects.In this study,we replaced the normal convolution in ASPP with a DSC unit to control the number of parameters.Experiments on the Potsdam and Vaihingen datasets demonstrated that our proposed model performs better than U-Net,DeepLabv3+,SwinUnet,and SegFormer.It is more sensitive to small-scale information and works better on small-scale features.In areas with darker images,all models are prone to misclassification.Providing calculation guidance for low-contrast areas will be considered in future work.

    Acknowledgement:The authors are highly thankful to the National Natural Science Foundation of China,to the Hainan Provincial Natural Science Foundation of China.Furthermore,we would also like to thank ISPRS for providing the remote sensing images.

    Funding Statement:This research was funded by the National Natural Science Foundation of China(Grant Number 62066013) and Hainan Provincial Natural Science Foundation of China (Grant Numbers 622RC674 and 2019RC182).

    Author Contributions:The authors confirm the following contributions to this paper:study conception and design:H.Xing,Q.Tong;data collection:K.Cao,Z.Zhu;writing—review&editing,analysis and interpretation of results:Q.Tong,H.Xing,M.Zhang,K.Cao;draft manuscript preparation:Q.Tong,Z.Zhu,K.Cao.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data presented in this study are available upon request from the corresponding author.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲,欧美,日韩| 女人十人毛片免费观看3o分钟| 亚洲欧美一区二区三区国产| 亚洲在线自拍视频| 欧美xxⅹ黑人| 国产精品人妻久久久影院| 欧美日韩视频高清一区二区三区二| 国产精品久久久久久精品电影小说 | 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 69人妻影院| 亚洲人与动物交配视频| 久久亚洲国产成人精品v| av网站免费在线观看视频 | 国产精品蜜桃在线观看| 免费av观看视频| ponron亚洲| 777米奇影视久久| videossex国产| 99热这里只有精品一区| 亚洲经典国产精华液单| 国产伦精品一区二区三区视频9| 日本爱情动作片www.在线观看| 久久精品国产亚洲网站| 美女被艹到高潮喷水动态| 欧美xxⅹ黑人| 色综合站精品国产| 国产精品一二三区在线看| 久久久久久久久中文| 成人毛片a级毛片在线播放| 黑人高潮一二区| 精品久久久精品久久久| 天堂网av新在线| 夜夜看夜夜爽夜夜摸| 免费av毛片视频| 午夜福利视频精品| 国产一区有黄有色的免费视频 | 久久97久久精品| 女人被狂操c到高潮| 亚洲精品一二三| 亚洲精品自拍成人| 国产高清有码在线观看视频| 99热这里只有精品一区| 肉色欧美久久久久久久蜜桃 | 国产亚洲精品久久久com| ponron亚洲| 欧美精品国产亚洲| 亚洲国产精品成人久久小说| 波多野结衣巨乳人妻| 精品人妻偷拍中文字幕| 18禁在线播放成人免费| 亚洲国产成人一精品久久久| 亚洲成人av在线免费| 精品国产三级普通话版| 男插女下体视频免费在线播放| 国产在视频线在精品| 国产高清不卡午夜福利| 午夜激情久久久久久久| av女优亚洲男人天堂| 精品国内亚洲2022精品成人| 纵有疾风起免费观看全集完整版 | 特大巨黑吊av在线直播| 中文字幕久久专区| 国产毛片a区久久久久| 国产有黄有色有爽视频| 边亲边吃奶的免费视频| 国产真实伦视频高清在线观看| 一级毛片 在线播放| 国产精品爽爽va在线观看网站| 卡戴珊不雅视频在线播放| 男人爽女人下面视频在线观看| 欧美成人精品欧美一级黄| 国产淫语在线视频| 亚洲av二区三区四区| 日韩一本色道免费dvd| 欧美bdsm另类| 午夜激情久久久久久久| 一个人观看的视频www高清免费观看| 国产精品熟女久久久久浪| 中文字幕免费在线视频6| 成人午夜精彩视频在线观看| 69av精品久久久久久| 久久97久久精品| 五月天丁香电影| 韩国高清视频一区二区三区| 亚洲天堂国产精品一区在线| 精品国内亚洲2022精品成人| av天堂中文字幕网| ponron亚洲| 久久99蜜桃精品久久| 搡女人真爽免费视频火全软件| 亚洲伊人久久精品综合| 国产欧美另类精品又又久久亚洲欧美| 一级片'在线观看视频| 在线 av 中文字幕| 超碰av人人做人人爽久久| 午夜免费激情av| 狠狠精品人妻久久久久久综合| 国产日韩欧美在线精品| 熟妇人妻久久中文字幕3abv| 精品少妇黑人巨大在线播放| 边亲边吃奶的免费视频| 国产精品av视频在线免费观看| 中国国产av一级| 日本一二三区视频观看| 欧美xxxx性猛交bbbb| 久久久精品免费免费高清| 国产老妇伦熟女老妇高清| 欧美xxxx性猛交bbbb| 国产精品一二三区在线看| 久热久热在线精品观看| av线在线观看网站| 国产精品不卡视频一区二区| 成人亚洲欧美一区二区av| 黄片无遮挡物在线观看| 欧美性猛交╳xxx乱大交人| 91久久精品国产一区二区成人| 久久久久精品性色| 狠狠精品人妻久久久久久综合| 成人综合一区亚洲| 免费大片黄手机在线观看| 免费观看av网站的网址| 亚洲精品日韩av片在线观看| 在线 av 中文字幕| 欧美 日韩 精品 国产| 久久久欧美国产精品| av.在线天堂| 久久久久精品久久久久真实原创| 亚洲精品国产成人久久av| 亚洲精品日本国产第一区| 伊人久久国产一区二区| 久久精品人妻少妇| 亚洲精品影视一区二区三区av| 亚洲人成网站高清观看| 日本与韩国留学比较| 女人十人毛片免费观看3o分钟| 天堂影院成人在线观看| 欧美精品一区二区大全| 91在线精品国自产拍蜜月| 亚洲久久久久久中文字幕| 欧美三级亚洲精品| 在线观看美女被高潮喷水网站| 精品久久久久久成人av| 亚洲精品自拍成人| 亚洲精品影视一区二区三区av| 午夜精品一区二区三区免费看| 亚洲精品自拍成人| 国模一区二区三区四区视频| 久久久久免费精品人妻一区二区| 美女国产视频在线观看| 精品人妻熟女av久视频| 女人久久www免费人成看片| 日韩三级伦理在线观看| 免费看日本二区| 精品国产一区二区三区久久久樱花 | 国产精品一二三区在线看| 白带黄色成豆腐渣| 又粗又硬又长又爽又黄的视频| 亚洲av国产av综合av卡| 免费观看性生交大片5| 伊人久久国产一区二区| 中文欧美无线码| 毛片一级片免费看久久久久| 日韩一区二区视频免费看| 国产真实伦视频高清在线观看| videossex国产| 美女被艹到高潮喷水动态| 日韩欧美精品v在线| 久久午夜福利片| 可以在线观看毛片的网站| 亚洲成人精品中文字幕电影| 国产老妇伦熟女老妇高清| 人妻一区二区av| 99久久精品热视频| 青春草国产在线视频| 久久久久久久久久久免费av| 激情五月婷婷亚洲| av黄色大香蕉| 亚洲欧美日韩无卡精品| 一本一本综合久久| 国产精品一区二区三区四区免费观看| 精品久久国产蜜桃| 高清在线视频一区二区三区| 男女边摸边吃奶| 搡老妇女老女人老熟妇| 久久久精品免费免费高清| 少妇丰满av| 在线 av 中文字幕| 免费av不卡在线播放| 国产精品日韩av在线免费观看| 欧美日韩亚洲高清精品| 22中文网久久字幕| 能在线免费看毛片的网站| 80岁老熟妇乱子伦牲交| 久久精品国产自在天天线| 大香蕉97超碰在线| 欧美激情久久久久久爽电影| 综合色av麻豆| 黄色欧美视频在线观看| 日本wwww免费看| 在线免费十八禁| 少妇熟女aⅴ在线视频| 欧美精品国产亚洲| 尤物成人国产欧美一区二区三区| 国产av在哪里看| 亚洲综合精品二区| 免费大片黄手机在线观看| 一级片'在线观看视频| 18禁动态无遮挡网站| 午夜福利在线观看免费完整高清在| 欧美性猛交╳xxx乱大交人| 亚洲天堂国产精品一区在线| 国产一区二区在线观看日韩| 免费黄色在线免费观看| 全区人妻精品视频| av卡一久久| 亚洲av中文字字幕乱码综合| 2021天堂中文幕一二区在线观| 国产男人的电影天堂91| 精品酒店卫生间| 热99在线观看视频| 春色校园在线视频观看| 免费观看精品视频网站| 天天躁夜夜躁狠狠久久av| xxx大片免费视频| 少妇人妻精品综合一区二区| 色综合色国产| 免费电影在线观看免费观看| 男人爽女人下面视频在线观看| 欧美日本视频| 日本-黄色视频高清免费观看| 国产成人福利小说| a级毛色黄片| 国内精品一区二区在线观看| 我的女老师完整版在线观看| 亚洲欧美中文字幕日韩二区| 国产伦理片在线播放av一区| 欧美精品一区二区大全| 亚洲综合色惰| 男人爽女人下面视频在线观看| 在线播放无遮挡| 婷婷色麻豆天堂久久| 日日撸夜夜添| 性插视频无遮挡在线免费观看| 国产精品久久久久久精品电影小说 | 国产精品国产三级专区第一集| 偷拍熟女少妇极品色| 亚洲国产成人一精品久久久| 国产一级毛片七仙女欲春2| 狂野欧美激情性xxxx在线观看| 女的被弄到高潮叫床怎么办| 特级一级黄色大片| 99热网站在线观看| 最近2019中文字幕mv第一页| 久久精品国产亚洲av涩爱| 亚洲怡红院男人天堂| 成人av在线播放网站| 熟妇人妻久久中文字幕3abv| 成人毛片60女人毛片免费| 97在线视频观看| 国产精品蜜桃在线观看| 我的老师免费观看完整版| 搡老妇女老女人老熟妇| 国产高清三级在线| 免费观看在线日韩| 高清av免费在线| 美女高潮的动态| 在线a可以看的网站| 激情五月婷婷亚洲| 亚洲久久久久久中文字幕| .国产精品久久| 国产精品一区二区三区四区久久| 美女高潮的动态| 国产熟女欧美一区二区| 激情 狠狠 欧美| 免费黄网站久久成人精品| 九九在线视频观看精品| 国产一区亚洲一区在线观看| 国产美女午夜福利| 99re6热这里在线精品视频| 久久精品夜夜夜夜夜久久蜜豆| 精品久久久精品久久久| 尤物成人国产欧美一区二区三区| 夫妻性生交免费视频一级片| 97超视频在线观看视频| 好男人在线观看高清免费视频| 天堂av国产一区二区熟女人妻| 精品人妻偷拍中文字幕| 少妇裸体淫交视频免费看高清| 99热这里只有精品一区| 亚洲av二区三区四区| 少妇被粗大猛烈的视频| 91精品伊人久久大香线蕉| 欧美极品一区二区三区四区| www.av在线官网国产| 免费播放大片免费观看视频在线观看| 日韩欧美精品免费久久| 老师上课跳d突然被开到最大视频| 国产永久视频网站| 青春草亚洲视频在线观看| 亚洲av电影不卡..在线观看| 寂寞人妻少妇视频99o| 少妇的逼好多水| 亚洲精品乱码久久久久久按摩| 日韩精品有码人妻一区| 亚洲av免费高清在线观看| 国产探花在线观看一区二区| freevideosex欧美| 搡老妇女老女人老熟妇| 久久精品国产亚洲av天美| 亚洲国产高清在线一区二区三| 国产黄a三级三级三级人| 三级国产精品片| 国产精品1区2区在线观看.| 国产精品日韩av在线免费观看| 欧美极品一区二区三区四区| 亚洲av日韩在线播放| 卡戴珊不雅视频在线播放| 亚洲乱码一区二区免费版| 国内精品美女久久久久久| 免费看美女性在线毛片视频| 乱码一卡2卡4卡精品| 亚洲欧美成人精品一区二区| 久久精品熟女亚洲av麻豆精品 | 亚洲精品,欧美精品| 18禁在线无遮挡免费观看视频| 80岁老熟妇乱子伦牲交| 亚洲怡红院男人天堂| 秋霞在线观看毛片| 一个人看的www免费观看视频| 99久久中文字幕三级久久日本| 国产视频内射| 在线天堂最新版资源| 美女xxoo啪啪120秒动态图| 国产亚洲av片在线观看秒播厂 | 亚洲国产精品成人综合色| 精品久久久噜噜| 日本色播在线视频| 麻豆成人午夜福利视频| 欧美变态另类bdsm刘玥| 日韩欧美精品v在线| 菩萨蛮人人尽说江南好唐韦庄| 青春草亚洲视频在线观看| 亚洲成人一二三区av| 你懂的网址亚洲精品在线观看| 国产视频首页在线观看| videossex国产| 国产人妻一区二区三区在| 中文资源天堂在线| 国产成人免费观看mmmm| 免费看av在线观看网站| 亚洲av日韩在线播放| 嫩草影院入口| 国产伦在线观看视频一区| 麻豆国产97在线/欧美| av在线老鸭窝| 国产av在哪里看| 国产淫片久久久久久久久| 国产精品99久久久久久久久| 男女国产视频网站| 观看免费一级毛片| 亚洲欧美成人综合另类久久久| 女人被狂操c到高潮| 男的添女的下面高潮视频| 好男人视频免费观看在线| 欧美激情国产日韩精品一区| 亚洲精品色激情综合| 亚洲一区高清亚洲精品| 日韩欧美 国产精品| 欧美日韩国产mv在线观看视频 | 成人午夜高清在线视频| av黄色大香蕉| 国产精品无大码| 好男人在线观看高清免费视频| 国内精品一区二区在线观看| 欧美性感艳星| 亚洲丝袜综合中文字幕| 国产伦精品一区二区三区视频9| 国产精品美女特级片免费视频播放器| 搞女人的毛片| 又粗又硬又长又爽又黄的视频| 91av网一区二区| 午夜老司机福利剧场| 特级一级黄色大片| 一级毛片 在线播放| 国产精品伦人一区二区| 中文资源天堂在线| av又黄又爽大尺度在线免费看| 成人鲁丝片一二三区免费| 午夜福利在线观看免费完整高清在| 69人妻影院| 中文字幕人妻熟人妻熟丝袜美| 深夜a级毛片| 亚洲人与动物交配视频| 色网站视频免费| 精品久久久久久久久av| 午夜福利成人在线免费观看| 亚洲精品久久久久久婷婷小说| 能在线免费看毛片的网站| 高清视频免费观看一区二区 | 成人av在线播放网站| 国产在线男女| 精品国产三级普通话版| 免费黄色在线免费观看| 搞女人的毛片| 黑人高潮一二区| 国产美女午夜福利| 国产白丝娇喘喷水9色精品| 日本猛色少妇xxxxx猛交久久| 18禁动态无遮挡网站| 三级国产精品欧美在线观看| 啦啦啦中文免费视频观看日本| 国产在视频线在精品| 网址你懂的国产日韩在线| 久久人人爽人人爽人人片va| 欧美日韩视频高清一区二区三区二| 欧美成人精品欧美一级黄| 搞女人的毛片| 美女国产视频在线观看| 亚洲成人一二三区av| 午夜日本视频在线| 日日摸夜夜添夜夜爱| 欧美变态另类bdsm刘玥| 亚洲国产欧美在线一区| 国产一级毛片在线| 两个人的视频大全免费| 欧美日韩视频高清一区二区三区二| 久久精品国产鲁丝片午夜精品| 2021天堂中文幕一二区在线观| 91精品伊人久久大香线蕉| 国产欧美日韩精品一区二区| 91aial.com中文字幕在线观看| 一区二区三区四区激情视频| 大又大粗又爽又黄少妇毛片口| 精品国产一区二区三区久久久樱花 | 中文精品一卡2卡3卡4更新| 久久草成人影院| 精品熟女少妇av免费看| 成人午夜高清在线视频| 亚洲美女搞黄在线观看| 人妻夜夜爽99麻豆av| 国精品久久久久久国模美| 简卡轻食公司| 国产精品久久久久久精品电影| 超碰97精品在线观看| 人妻夜夜爽99麻豆av| 成人无遮挡网站| 国产一区二区三区综合在线观看 | 91在线精品国自产拍蜜月| 成人欧美大片| 国产欧美另类精品又又久久亚洲欧美| 国产v大片淫在线免费观看| freevideosex欧美| 国产亚洲精品久久久com| 国产男女超爽视频在线观看| 综合色丁香网| 人人妻人人看人人澡| 午夜激情久久久久久久| 熟妇人妻不卡中文字幕| 国产精品无大码| 国产精品国产三级国产专区5o| 日韩欧美 国产精品| 欧美激情在线99| 日韩在线高清观看一区二区三区| av又黄又爽大尺度在线免费看| 狠狠精品人妻久久久久久综合| 身体一侧抽搐| 精品久久久久久久末码| 欧美人与善性xxx| 国产一区亚洲一区在线观看| videos熟女内射| 亚洲av.av天堂| 欧美高清性xxxxhd video| 国产午夜精品久久久久久一区二区三区| 亚洲综合精品二区| 久久久久国产网址| 久久久久久国产a免费观看| 久久久久久久大尺度免费视频| 亚洲自拍偷在线| 99热这里只有是精品在线观看| 亚洲精品aⅴ在线观看| 欧美xxxx黑人xx丫x性爽| 日本熟妇午夜| 乱人视频在线观看| 国产色婷婷99| 菩萨蛮人人尽说江南好唐韦庄| 晚上一个人看的免费电影| 汤姆久久久久久久影院中文字幕 | 久久草成人影院| 国产乱人视频| 99久国产av精品国产电影| 人人妻人人澡人人爽人人夜夜 | 中文字幕制服av| 亚洲成人精品中文字幕电影| 在线免费观看不下载黄p国产| 两个人的视频大全免费| 午夜福利视频1000在线观看| 国产精品1区2区在线观看.| 一二三四中文在线观看免费高清| 伦精品一区二区三区| 激情 狠狠 欧美| 亚洲丝袜综合中文字幕| 国产成人精品婷婷| 精品亚洲乱码少妇综合久久| 精华霜和精华液先用哪个| 99热这里只有是精品在线观看| 婷婷色麻豆天堂久久| 水蜜桃什么品种好| 亚洲第一区二区三区不卡| 国产综合精华液| 日韩视频在线欧美| 国产国拍精品亚洲av在线观看| 伊人久久国产一区二区| 婷婷六月久久综合丁香| 中文字幕亚洲精品专区| 少妇丰满av| 丝瓜视频免费看黄片| 亚洲美女视频黄频| 尤物成人国产欧美一区二区三区| www.色视频.com| 国产在线一区二区三区精| 国产伦精品一区二区三区视频9| 久久精品国产亚洲av天美| 亚洲美女视频黄频| 亚洲图色成人| 亚洲综合色惰| 亚洲人成网站在线播| 亚洲成人中文字幕在线播放| 美女cb高潮喷水在线观看| 一个人观看的视频www高清免费观看| av线在线观看网站| 国产伦精品一区二区三区视频9| 内地一区二区视频在线| 亚洲欧美成人综合另类久久久| 亚洲精品456在线播放app| 97人妻精品一区二区三区麻豆| 嘟嘟电影网在线观看| 美女大奶头视频| 免费看日本二区| 婷婷色综合大香蕉| 欧美丝袜亚洲另类| 成人毛片60女人毛片免费| 国产精品嫩草影院av在线观看| 亚洲电影在线观看av| 精品久久久精品久久久| 伦理电影大哥的女人| 一区二区三区四区激情视频| www.av在线官网国产| 国产免费视频播放在线视频 | 一级毛片 在线播放| videos熟女内射| 亚州av有码| 成人一区二区视频在线观看| 淫秽高清视频在线观看| 人妻一区二区av| 国产av国产精品国产| 亚洲色图av天堂| 一区二区三区乱码不卡18| 高清午夜精品一区二区三区| 毛片女人毛片| 精品一区二区三区人妻视频| 久99久视频精品免费| 国产亚洲5aaaaa淫片| 国产精品久久久久久久久免| 国内揄拍国产精品人妻在线| 天堂影院成人在线观看| 精品午夜福利在线看| 国产伦精品一区二区三区四那| 成人漫画全彩无遮挡| 久久鲁丝午夜福利片| 毛片一级片免费看久久久久| av专区在线播放| av在线天堂中文字幕| 岛国毛片在线播放| 最后的刺客免费高清国语| 国产女主播在线喷水免费视频网站 | 成人av在线播放网站| 国产黄频视频在线观看| 亚洲久久久久久中文字幕| 久久99热这里只有精品18| 最近最新中文字幕大全电影3| 国产成年人精品一区二区| 精品一区二区三卡| 少妇的逼好多水| 国产免费又黄又爽又色| 狂野欧美激情性xxxx在线观看| 天天一区二区日本电影三级| 女人十人毛片免费观看3o分钟| 麻豆乱淫一区二区| 街头女战士在线观看网站| 九九爱精品视频在线观看| 亚洲综合色惰| 亚洲在线自拍视频| 久久久久久久午夜电影| 麻豆成人av视频| 少妇猛男粗大的猛烈进出视频 | www.av在线官网国产| 精品久久久久久久末码| 精品一区二区三区人妻视频| 亚洲欧美一区二区三区国产| 久久99蜜桃精品久久| 久久精品久久久久久久性| 日韩精品有码人妻一区| 精品久久久噜噜| or卡值多少钱| 女人久久www免费人成看片| 2021少妇久久久久久久久久久| 亚洲精品视频女| 久久久久久久大尺度免费视频| 色吧在线观看| 中文字幕av在线有码专区| 久久精品久久精品一区二区三区| 人体艺术视频欧美日本| 亚洲国产色片| 只有这里有精品99| 成年av动漫网址|