• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Rail Surface Defect Detection Based on Improved UPerNet and Connected Component Analysis

    2023-12-12 15:50:54YongzhiMinJiafengLiandYaxingLi
    Computers Materials&Continua 2023年10期

    Yongzhi Min,Jiafeng Li and Yaxing Li

    1School of Automation and Electrical Engineering,Lanzhou Jiaotong University,Lanzhou,730070,China

    2Gansu Provincial Engineering Research Center for Artificial Intelligence and Graphics&Image,Lanzhou Jiaotong University,Lanzhou,730070,China

    3State Grid Gansu Electric Power,Baiyin Power Supply Company,Baiyin,730900,China

    ABSTRACT To guarantee the safety of railway operations,the swiftdetection of rail surface defects becomes imperative.Traditional methods of manual inspection and conventional nondestructive testing prove inefficient,especially when scaling to extensive railway networks.Moreover,the unpredictable and intricate nature of defect edge shapes further complicates detection efforts.Addressing these challenges,this paper introduces an enhanced Unified Perceptual Parsing for Scene Understanding Network(UPerNet)tailored for rail surface defect detection.Notably,the Swin Transformer Tiny version (Swin-T) network,underpinned by the Transformer architecture,is employed for adept feature extraction.This approach capitalizes on the global information present in the image and sidesteps the issue of inductive preference.The model’s efficiency is further amplified by the windowbased self-attention,which minimizes the model’s parameter count.We implement the cross-GPU synchronized batch normalization(SyncBN)for gradient optimization and integrate the Lovász-hinge loss function to leverage pixel dependency relationships.Experimental evaluations underscore the efficacy of our improved UPerNet,with results demonstrating Pixel Accuracy(PA)scores of 91.39%and 93.35%,Intersection over Union(IoU)values of 83.69%and 87.58%,Dice Coefficients of 91.12%and 93.38%,and Precision metrics of 90.85%and 93.41%across two distinct datasets.An increment in detection accuracy was discernible.For further practical applicability,we deploy semantic segmentation of rail surface defects,leveraging connected component processing techniques to distinguish varied defects within the same frame.By computing the actual defect length and area,our deep learning methodology presents results that offer intuitive insights for railway maintenance professionals.

    KEYWORDS Rail surface defects;connected component analysis;Transformer;UPerNet

    1 Introduction

    In the intricate wheel-rail system,the rail serves dual pivotal roles:it bears the vehicular vertical load and facilitates guidance.The rail head,consequent to this,encounters longitudinal and transverse forces,engendering vibrations throughout the rail system.Such perturbations can lead to detrimental surface defects on the rail.Notably,these imperfections intensify the rail’s vertical vibrations,transmitting these oscillations to the locomotive.This amplification can heighten the potential for resonance between the vehicle and its assorted elastic components.Concurrently,an increase in the rail’s torque can be observed,undermining the wheel’s directional function and jeopardizing the vehicle’s lateral stability.Hence,prompt detection of these rail surface defects becomes imperative[1].

    The vast expanse of the railway network,coupled with a multifarious natural environment and dense departures,renders manual inspection an inefficacious solution.However,the advancement in machine vision coupled with the burgeoning field of deep learning has catapulted non-contact,imagebased defect detection methods into a focal research area.Traditional visual techniques predominantly utilize texture and rudimentary grayscale attributes to discern and pinpoint target zones in unlabeled rail samples.Classic methodologies employ a gradient histogram in tandem with support vector machines,realizing rail surface defect detection.This can be further amalgamated with multiple detectors within a multi-task learning paradigm to enhance detection prowess[2].A notable study by Li et al.[3]has accentuated defect contrasts via local normalization and defect proportion limitation filtering.Their approach employed a grayscale projection contour algorithm,displaying robustness to noise,with the added advantage of swift detection rates.Techniques encompassing the internal point hollowing algorithm,coupled with the chain code tracking algorithm,have demonstrated potential in capturing defect contour data,thereby enabling automated detection of rail surface defect regions [4].In another significant research [5],differentiation of the rail image post-inverse Perona-Malik(P-M)diffusion from the prototypical image was undertaken to minimize environmental interferences.Following this,the defect region was segmented through an edge feature-based filtering algorithm.To further refine the segmentation threshold of rail defects,a novel Otsu method integrating target variance weighting was introduced in [6].Some researchers have applied Laplace low-pass filtering to rail imagery,subsequently differentiating it from the original image.This was followed by morphological feature extraction for defect detection.On average,the accuracy of this methodology was found to be 85.3% [7].Yaman et al.[8] adopted a distinctive approach by extracting feature types from the variance signal of rail imagery,subsequently ascertaining the fault type via fuzzy logic.In light of the insensitivity of defect detection to the rail’s transverse luminosity alterations,a defect detection framework was formulated.This system,underpinned by morphological operators,employed elongated,slender structural elements aligned parallel to the rail’s longitudinal axis for image processing,effectively curtailing false positives[9].Confronting the challenges inherent in rail detection,such as intricate conditions and unequal rail luminescence,a region extraction algorithm hinged on vertical projection and grayscale differentiation was conceptualized.Furthermore,an enhanced rapid robust Gaussian mixture model anchored on the Markov random field was designed,aiming for swift and precise surface defect segmentation[10].Academic discourse has also illuminated a maximum entropy threshold segmentation technique.This method,rooted in the principle where the background entropy remains invariant in low gray domains,while the target entropy fluctuates in accordance with the gray scale,has shown promise in attaining optimal thresholds for demarcating rail surface defects[11].It is crucial to note,however,that while these methods have advanced defect object detection considerably,they demand scene-specific modifications.Thus,their adaptability and real-time performance warrant further enhancement.

    Deep learning methodologies employ convolutional neural networks(CNNs)to represent defects in rail images as abstract digital features for enhanced detection.The prevailing research predominantly capitalizes on object detection and semantic segmentation networks,epitomized by their rapid processing capabilities and obviation of manual feature setting.Certain approaches harness CNNs directly to categorize rail surface defects for detection intentions.For instance,a study delineated in [12] contrived a CNN comprised of 2 convolution layers,2 pooling layers,and a fully connected layer,expressly for rail surface defect detection;a regularization technique was incorporated during the training phase to augment the recognition rate.Faghih-Roohi et al.[13],in their exploration,classified rail defect images into six distinct categories and embarked on training three disparate structures of deep CNNs via the small batch gradient descent approach.Their findings underscored the superior classification performance of more profound network structures,with heightened accuracy when the rectified linear unit(ReLU)was integrated into training.Turning to detection methodologies based on semantic segmentation networks,a study referenced in[14]conceived a 59-layer semantic segmentation network rooted in the SegNet architecture to pinpoint rail surface anomalies.When juxtaposed against the Otsu segmentation and manual threshold segmentation techniques,this semantic segmentation network triumphed with the preeminent detection rate.Further,a study in[15]introduced a lightweight rail surface defect detection network that synergized MobileNetV2 with the You Only Look Once version3(YOLOv3)mechanism.By supplanting Darknet53 with MobileNetV2 for backbone extraction,and concurrently designing a multi-scale predictor inspired by target regression,an inference speed culminating at 87.40%mean average precision(mAP)and 60 frames per second(FPS)was realized.Jin et al.[16] amalgamated semantic segmentation with object detection techniques within their detection framework,integrating Dropout with Xception—a backbone network of DeepLabV3+.To counteract the pixel imbalance between the foreground and background,YOLO was employed to pinpoint defects and deduce foreground and background weighting coefficients,thereby furnishing attention loss for the segmentation model.Dilated convolution emerges as a prevalent strategy to amplify the receptive field.A paper cited in[17]devised a lightweight semantic segmentation model rooted in a cascading self-encoding network for defect detection in rails and crafted a compact CNN specifically for defect classification.This methodology flourished both in detection precision and temporal efficiency.Luo et al.’s works [18,19] incorporated dilated convolution into the Faster Regions with CNN Features(Faster R-CNN)and established parallel convolution channels,each with unique expansion coefficients catering to distinct defect types.This innovation bolstered the detection precision of nuanced defects.Subsequent investigations leveraged the Cascade Regions with CNN Features (Cascade R-CNN) for rail surface defect detection.Addressing challenges like the skewed distribution of sample IoU and misalignment between regions of interest and feature maps engendered by rounding quantization,and the inaccurate prognostication of regression boundaries,various tools such as IoU balanced sampling,region of interest aligns(RoIAlign),and complete intersection over union(CIoU)loss were assimilated to hone the average precision(AP)to a staggering 98.75%.While unsupervised methodologies have also been ventured into for rail surface defect detection,their prevalence remains limited.A study in[20],for example,employed scale normalized cross fast flow for probability distribution estimation,which facilitated the derivation of an image’s standard Gaussian distribution.Anomaly scores were subsequently discerned based on the proximity to the distribution’s epicenter,serving the dual purpose of detection and defect localization.

    Currently,the rail surface defect detection employing image methodologies confronts several challenges: (a) The discernible information present within rail images is somewhat circumscribed,leading to diminished efficiency of methods grounded on texture and grayscale features.(b) The proactive maintenance interventions by the concerned departments hinder the effective accumulation of a voluminous dataset of rail defect images,thereby compromising the efficacy of supervised deep learning.(c)Although a plethora of rail inspection vehicles are competent in detecting fastener defects,the image data they procure for rail damage differs significantly from conventional images,making them unsuitable for direct utilization.(d) Rail surface anomalies predominantly manifest in two distinct forms:ripple and discrete defects.The emergence of discrete defects is characterized by their unpredictability,arbitrariness,and multi-scale dimensions,devoid of any recurring patterns,which intensifies the challenges associated with defect detection [21].(e) The intricate edge configurations of rail surface defects pose considerable challenges to semantic segmentation algorithms.(f) The detection outcomes yielded by deep learning approaches are arduous to correlate directly with the specific parameters delineating the defects.

    In light of these challenges,an enhanced UPerNet semantic segmentation network was introduced specifically for rail defect detection.Concurrently,a Swin-T network integrating the Transformer architecture was deployed for the extraction of defect features,which capitalized on the holistic information embedded within the images while circumventing the issue of inductive bias.The rule window self-attention inherent in Swin-T mitigates the model’s parameter load.Additionally,the cross-GPU synchronized batch normalization technique was harnessed for gradient refinement,with Lovász-hinge being employed as the loss function,facilitating the leveraging of inter-pixel dependency relationships.Conclusively,grounded on the semantic segmentation imagery of rail surface anomalies,the distinct defect regions were identified via connected component analysis,enabling the computation of the genuine length and area of the defects.

    The contributions of this research can be dichotomized as:

    (a)The proposition of a rail surface defect detection algorithm founded on semantic segmentation.Within this framework,the Swin-T,encompassing the Transformer architecture,is incorporated,amplifying the capability of the segmentation network in the extraction of rail surface defect features,and bolstering the defect segmentation quality.Notwithstanding the pervasive application of Transformers across a myriad of visual tasks—attributed to their prowess in assimilating global context information—their deployment for this specific endeavor remains an exception.

    (b)Pivoting from the semantic segmentation representations of rail surface anomalies,connected component analysis is integrated,which adeptly bridges the detection outcomes of the deep learning modus operandi with the intricate parameters defining the defects.Such an approach aids railway maintenance personnel in intuitively grasping the defect metrics,enabling them to orchestrate targeted remediation measures.

    2 Related Work

    2.1 Transformer

    The Transformer architecture,a manifestation of the encoder-decoder network paradigm,is predominantly anchored on the attention mechanism[22].Dispensing with recursion and convolution,it rectifies the inherent sluggishness that plagues the training of Recurrent Neural Networks(RNN)and has demonstrated its proficiency in machine translation endeavors.Models grounded in the Transformer architecture,such as Bidirectional Encoder Representation from Transformers (BERT) [23]and Generative Pre-train Transformer(GPT)[24],have been recognized for their stellar performance in the realm of natural language processing.For myriad computer vision tasks,the prototypical Transformer encoder model is employed,serving as a novel feature extraction apparatus.The Vision Transformer (ViT) [25],designed for image classification,segments images into standardized image blocks.These embedded blocks are subsequently input into the Transformer encoder,and post Multilayer Perceptron(MLP)classification processing,it garners commendable results,rivaling toptier convolutional networks.The DEtection TRansformer(DETR)[26]amalgamates both CNN and Transformer architectures to present an end-to-end object detection framework,reframing object detection as a images-to-sets prediction conundrum.By leveraging the self-attention mechanism of the Transformer,it simulates all pairwise interactions amidst elements in a sequence,rendering it more conducive for the intricate constraints intrinsic to set prediction.Chen et al.[27] devised the Image Processing Transformer (IPT),a model sculpted on the Transformer architecture,adept at addressing image processing challenges,excelling in tasks like image super-resolution,restoration,and denoising across a plethora of marred images from ImageNet.As per Reference [28],a semisupervised video segmentation blueprint was proffered,which incorporated the Transformer at its bottleneck layer to further distill contextual data,thereby enhancing the precision of endoscopic video segmentation in scenarios characterized by scarce annotated data.Despite the widespread adoption and commendation of the Transformer structure across a variety of visual tasks,attributed to its prowess in gleaning global contextual features,its deployment in the domain of rail surface defect detection remains limited.

    2.2 Attention Mechanism

    The genesis of the attention mechanism can be traced back to studies delving into human visual cognition.Given the constraints imposed by information processing bottlenecks,humans innately zero in on certain segments of the information spectrum,while disregarding the remainder.Emulating this intrinsic human trait,the attention mechanism in deep learning serves as an intuitive resource distribution conduit,reallocating weights contingent on the perceived significance of entities.

    The Transformer intricately weaves multiple self-attention layers,culminating in multi-head selfattention(MSA),with an aim to enhance network efficiency,as depicted in Fig.1.The self-attention layer transmutes the input vector into three distinct entities:the query vector(q),the key vector(k),and the value vector(v),subsequently coalescing them into three matrices:Q,K,andV.Attention weights are ascertained through a liaison betweenQandK,and subsequently,the comprehensive weight and output are procured via their interaction withV.Vectors deemed of higher significance are accorded heightened attention in subsequent layers.The self-attention dynamic across disparate input vectors is delineated in formula(1),wheredkrepresents the dimension pertinent toK.

    Figure 1:The principle of multi-head self-attention mechanism

    The concept of multi-head self-attention gives rise to the simultaneous execution of multiple selfattention operations on features within the network.These individual operations are subsequently amalgamated to produce the final output,enabling the network to encapsulate a richer spectrum of feature information.This process is mathematically articulated in formula(2).

    where,headiis the result of each self-attention action,as shown in formula(3),WQi,WKiandWViare all projection weight matrices.

    3 Proposed Methodology

    The UPerNet[29],a semantic segmentation network,is predicated upon both the Pyramid Pooling Module(PPM)and the Feature Pyramid Networks(FPN).Drawing upon four distinct feature layers as elucidated by ResNet,it meticulously weaves together multi-scale information derived from various stages,ensuring a minimized loss of boundary information.This is a marked improvement over the Pyramid Scene Parsing Network(PSPNet)[30],another network subscribing to the FPN framework.Yet,the UPerNet’s reliance on CNN for feature extraction manifested an inherent inductive bias,hindering the holistic utilization of global information.Moreover,when training is undertaken on a solitary GPU with a comparably smaller batch,the application of BN (Batch Normalization)tends to compromise the model’s convergence efficacy.Further,the employment of the cross-entropy loss function overlooks the intricate dependencies present amongst adjacent pixels.In light of these challenges,enhancements were made to the UPerNet,with its overarching structure delineated in Fig.2.The encoder network,seeking refuge in Swin-T [31] for feature extraction,collaborates with the decoder network,which harnesses both PPM and FPN.This ensures a seamless projection of discernible features into the pixel space,post-upsampling.The final ensemble of feature maps,stemming from each stage of the feature extraction network,is encapsulated as {S2,S3,S4,S5}.Concurrently,the feature mapping borne out of the FPN output is represented as {P2,P3,P4,P5},with P5 also succeeding PPM directly.

    Figure 2:The overall structure of the method in this paper

    3.1 Backbone Network

    While CNNs have demonstrated proficiency across various image processing tasks,the inherent locality of convolutional operations and the inductive bias stemming from weight sharing impose constraints on long-distance dependencies.Such limitations curtail the ability of CNNs to glean comprehensive insights from an entire image and to facilitate information exchange beyond the confines of the convolutional kernel.In stark contrast,the self-attention mechanism inherent in the Transformer can adaptively modify its receptive domain,thereby furnishing long-range dependencies and circumventing the inductive bias endemic to CNNs.In this study,the Swin-T architecture is employed as the backbone network,facilitating the extraction of expansive global information and more nuanced features.The strategic confinement of attention computation to non-overlapping local windows ensures the computational complexity remains linear in relation to image size,thereby optimizing computational efficiency.A schematic representation of this network structure can be perused in Fig.3.

    Figure 3:Network structure of Swin-T

    The entire network adopts a stratified approach.Initially,an input image of dimensionsH×W× 3 is partitioned into (H/4) × (W/4) non-overlapping patches via the patch-splitting module,subsequently flattening each patch into a 48-dimensional token vector.This is followed by the linear embedding layer projecting the tensor of dimensions(H/4)×(W/4)×48 to an arbitrary dimensionC,culminating in a linear mapping of dimension(H/4)×(W/4)×C.These mapped tokens are then channelled into the dual Swin Transformer module,which ensures the token count remains consistent at(H/4)×(W/4)both at the input and output.The inaugural Swin Transformer module,in conjunction with the linear embedding layer,is designated as the primary layer.Thereafter,the patch merging layer collaborates with the Swin Transformer module across subsequent layers to elucidate feature descriptors across varied scales.Swin Transformer diverges from traditional architectures like ResNet,opting to segregate the input image via diminutive windows,consequently parsing it into equitably sized blocks termed local windows.Within each of these local windows,feature extraction ensues,thereby permitting both self-attention and localized convolutional operations to ascertain feature information across diverse scales and positions.numerous Swin Transformer Blocks(analogous to the basic blocks in ResNet)are harnessed for feature extraction across Stages 2–4.Eschewing the simplistic stacking of convolutional layers as observed in ResNet,Swin Transformer adopts a hierarchical group convolution to optimize feature amalgamation.Furthermore,Swin Transformer facilitates inter-window information exchange,enhancing the context-dependence of features.As the network becomes progressively intricate,every layer metamorphoses the tensor’s dimension,culminating in a stratified representation.Pertinently,each patch boasts a size of 4,each self-attention window spans 7,and the network depth across the four blocks is catalogued as[2,2,6,2].In summation,the multifaceted stages of Swin Transformer utilize local windows,hierarchical group convolution,and inter-window connections to glean feature insights across an array of positions and scales.This epitomizes Swin Transformer’s adeptness in tailoring its feature extraction capabilities to cater to diverse tasks and datasets.

    The Swin Transformer Block is composed of Layer Normalization (LN),a multi-head selfattention module,a shortcut connection,and anMLPfeaturing a Gaussian Error Linear Unit(GELU)activation function.A representation of two consecutively arranged Swin Transformer Blocks is depicted in Fig.4.Within the design of the multi-head self-attention module,the self-attention mechanism is intricately integrated into the window,encompassing two distinct types: the windowbased self-attention (W-MSA) and the shifted window-based self-attention (SW-MSA).TheWMSAmode partitions the image into multiple windows and executes self-attention computations individually for each window.Given an image fragmented intoh×wpatches,where each window comprisesM×Mpatches,the computational complexities ofMSAandW-MSAare elucidated in formulas(4) and (5),respectively.A noteworthy reduction in computational complexity is achieved throughW-MSA.Owing to the fact that the patch count within each window is significantly inferior to the total patch count of the image,the computational complexity associated withW-MSAexhibits linearity with respect to the image’s dimensions.

    SW-MSA,on the other hand,incorporates a shifted window approach,facilitating inter-window connections.This not only augments information transfer across different windows but also amplifies the receptive field,thereby bolstering the model’s representational prowess.The interrelationships embedded within the two serially arranged Swin Transformer Blocks are expounded in formula(6),where ?zlandzlrepresent the features output by the two types of window self-attention and the features output by theMLP,respectively.

    Figure 4:Two swin Transformer blocks in series

    3.2 Synchronized Batch Normalization

    In traditional applications,when batch normalization (BN) serves as the chosen normalization technique,the inputx={x1,x2,x3,x4} is segmented into multiple portions during each iteration of network training.These portions undergo forward and backward propagation processes,with the gradients being computed on disparate GPUs.Subsequently,a succeeding iteration is commenced post the amalgamation of gradients and parameters.This fundamental operation is visualized in Fig.5.

    Figure 5:Working principle of BN

    Due to each GPU processing data individually during both forward and backward propagations,the BN operation is independently executed on every GPU.Intrinsically,the samples normalized via BN are confined to their respective GPU,mirroring a reduction in the batch_size.A diminished batch_size can lead to the calculated mean and variance failing to capture the holistic data distribution,as a smaller batch_size introduces heightened variability.Notably,in the realm of deep learning,an expansive batch_size typically stabilizes the training trajectory.Cross-GPU synchronized batch normalization(SyncBN)retrieves the universal meanμand varianceσ2during the forward propagation phase,and computes the overarching gradient throughout the backward propagation.When deducing the global gradient in the backward propagation,the global BN statistics can be employed for input normalization,which invariably refines the accuracy of these statistics,as depicted in Fig.6.

    SyncBN ensures that the input data distribution across every network layer remains comparably consistent,permitting the adoption of larger learning rates.This facilitates the model’s diminished sensitivity towards network parameters,attenuates its reliance on initialization,simplifies the parameter tuning procedure,and promotes stable network learning.By leveraging normalized data subsequent to the activation function,SyncBN sustains gradients at relatively ample values,averting gradient vanishing or explosive behaviors.Such synchronized normalization has been effectively incorporated in applications like power grid device image anomaly detection[32]and apparel image analysis[33],facilitating seamless synchronization of network parameters across GPUs.

    Figure 6:Working principle of SyncBN

    Within each GPU,both the summations Σxiand Σxi2,are computed.Following this,a synchronized summation is executed to deduce the global variance and mean.Such a method allows the entire calculation to be finalized with a singular synchronization.This effectively reduces the number of cross-GPU synchronizations from two to just one,enhancing the speed of synchronized normalization,as articulated in formulas(7)and(8).Herein,xirepresents the sample point,andmis the cumulative number of sample points spanning multiple GPUs.

    3.3 Loss Function

    In scenarios where cross-entropy functions as the segmentation loss,certain complications arise.Particularly,when the quantity of foreground pixels is significantly lower than that of background pixels,the inherent competition mechanism of the cross-entropy loss function results in a predominance of the background pixels.Such a skew can severely bias the optimization trajectory of the model towards the background,proving detrimental,especially for segmenting minute rail surface defects.Additionally,cross-entropy loss,given its pixel-by-pixel computation nature,overlooks interpixel relationships.It is crucial to note that images inherently contain significant dependencies between pixels,which often encapsulate vital information about the structure of the subject.Relying solely on cross-entropy loss could lead to segmentation outputs that harbor imprecise,ambiguous predictions.

    Weeks later, when the real seeds began to break through the soil, the boy became entranced with the tiny seedlings6. He spent many afternoons helping7 Granddad water and hoe and watch them grow. And when the first baby vegetables were harvested, he liked them after all.

    To address these challenges,this study integrates the Lovász-hinge loss [34].This loss function inherently embodies the dependencies existing between pixels situated at varied positions within an image and further takes into account the disparity between background and target pixels.The computation associated with Lovász-hinge is inherently linked with the Jaccard Index (IoU).The pertinent loss function is illustrated in formula(9),wherein ?yrepresents the model’s prediction vector,y?denotes the ground truth,andcsymbolizes the object class.The set constituting the mispredicted pixels for classcis articulated in formula(10).Further,formula(9) can be reconceptualized as a function of the misprediction set,as delineated in formula(11).

    In the use of the Jaccard loss,it is mandated that the prediction result be discrete;a continuous prediction result cannot be derived post-discretization.Given that the Jaccard loss operates as a submodule function,its input space can be transformed from discrete {0,1}pto continuous Rpthrough the Lovász extension.Furthermore,this transformed value remains consistent with the output value of the original function on {0,1}p.In addressing binary segmentation challenges such as rail surface defect segmentation,this study employs the continuous and derivable Lovász-hinge,which is attained after the smooth extension of Jaccard loss,as the loss function for optimal model parameterization.The Lovász-hinge loss has also found applications in the first arrival picking problem in seismic signal processing.In this context,the first arrival picking problem is reframed as a contour detection challenge,with the aim of enhancing the accuracy of first arrival picking[35].Kerola et al.[36] introduced a groundbreaking representation learning methodology tailored for panoptic segmentation.They further presented a streamlined architecture and loss mechanism to master pixel-wise embeddings,effectively representing instances,object categories,and stuff classes.This paved the way for a unified embedding-based single-shot panoptic segmentation.The entity ΔJcis reconceptualized as a suite of functions that mispredictm,with the mathematical representation of the incorrect prediction definition illustrated in formula(12).

    The continuous linear distribution function after Lovász extension of ΔJcis shown in formula(13).

    The mispredictionmiis replaced by the hinge loss of the input imagex,as shown in formula(14),whereF(x)represents the network output.

    g(m)is obtained by taking the derivative ofm,as shown in formula(15).

    To sum up,the loss function used in this paper is shown in formula(16).

    4 Experiment

    Experimental verification was undertaken on two datasets in this study.The initial dataset was procured from the transportation railway line of Baiyin Nonferrous Group Co.,Ltd.,where the primary defect type identified was shelling defects.Given that the background image information of rail ties and ballast was redundant for detection,it was consistently transformed into 160 × 370 pixels images encompassing only the rail surface area.A total of 914 defect images were retained post-processing.This dataset was christened the Rail-BY dataset.The subsequent dataset was dubbed the Rail Surface Discrete Defects(RSDDs),curated by researchers from Beijing Jiaotong University[37].Owing to the proactive intervention of the railway maintenance department,a scarcity of rail surface defect samples was observed.An expansion of defect images was executed through strategies like rotation,mirroring,filtering,denoising,and leveraging deep convolutional generative adversarial networks(DCGAN).The augmented Rail-BY dataset encompassed 9570 defect images:8134 in the training set,957 for validation,and 479 for testing.The RSDDs dataset included 6950 defect images,segmented into 5907 for training,695 for validation,and 348 for testing.Representative samples are depicted in Figs.7 and 8,with additional expanded samples in Fig.9.

    Figure 7:A partial sample of the Rail-BY dataset

    Figure 8:A partial sample of the RSDDs dataset

    The apparatuses harnessed for experimentation comprised an Intel Core i9-10920X CPU@3.50 GHz,RAM of 128 GB,dual NVIDIA GeForce RTX 3090 GPU,underpinned by Python 3.8 and PyTorch 1.8.1.During the experiments,a benchmark of 160000 iterations was established,utilizing the AdamW optimizer.Parameters were set atβ1=0.9,β2=0.999,with a learning rate of 0.00006,batch_size of 2,and weight_decay set to 0.01.Loss trajectory for the validation set across the Rail-BY and RSDDs datasets is illustrated in Fig.10.A stability in training loss post the 160000 iterations indicates an optimal training outcome.

    To ascertain the efficacy of the enhanced methodology presented in this study,a comparative examination of the segmentation outcomes of various modules on the two datasets was performed.Table 1 delineates the segmentation outcomes for diverse modules.The foundational backbone network utilized was ResNet50,with normalization facilitated via BN,and the loss function being the cross-entropy loss function.An examination of Table 1 reveals that solely substituting ResNet50 with Swin-T ameliorated PA,IoU,Dice,and Precision by 8.06%,11.42%,7.3%,and 9.52%,respectively on the Rail-BY dataset.Conversely,on the RSDDs dataset,improvements were noted at 3.48%for PA,4.45% for IoU,2.59% for Dice,and 1.7% for Precision.A comprehensive enhancement in UPerNet’s segmentation accuracy was observed,with the most pronounced improvement reflected in the original UPerNet.Additionally,the incorporation of Swin-T for feature extraction culminated in a reduction of Params by 6.55 M.The backbone network grounded in the Transformer architecture decidedly surpassed CNN-based networks in terms of performance.The employment of the SyncBN normalization technique and the Lovász-hinge loss function predominantly augmented PA.Given that these don’t pertain to network structure modifications,their impact on network complexity was minimal.

    Table 1:Comparison of segmentation results of different modules in improved UPerNet

    Table 2 presents the training outcomes of the refined UPerNet network juxtaposed against other prevailing semantic segmentation networks across both datasets.The methodology advanced in this study manifested commendable outcomes on both datasets.When pitted against the high-performing DeepLabV3+,the Rail-BY dataset witnessed enhancements of 6.88%in PA,6.83%in IoU,and 4.2%in Dice.Precision escalated by 1.38%,whereas on the RSDDs dataset,PA surged by 1.13%,IoU by 0.64%,Dice by 0.37%,but Precision experienced a decrement of 0.41%.The original UPerNet network recorded an FPS of 20.1 Hz on the Rail-BY dataset and 20.2 Hz on the RSDDs dataset.However,the FPS of the improved UPerNet on the datasets diminished by 2.2 Hz and 2.3 Hz,respectively,rendering its detection speed performance relatively moderate.Notwithstanding,while the parameter count of the enhanced method presented herein was reduced relative to the original UPerNet model,it remains considerably expansive when compared to other mainstream semantic segmentation techniques.A visual comparison between the methodology advanced in this study and the outcomes from other segmentation networks is portrayed in Fig.11.

    Table 2:Comparison of defect segmentation results between proposed method and other networks

    Figure 11:(Continued)

    Figure 11:Comparison of visualization results between proposed method and other semantic segmentation methods

    5 Defects Parameter Calculation

    Deep learning detection algorithms prevalent today are predominantly employed to ascertain the presence of rail defects,yet they seldom incorporate specific parameters associated with these defects.Consequently,in this work,the actual length,width,and area of rail surface defects are computed based on semantic segmentation.Such a computation aids railway maintenance personnel by providing an intuitive comprehension of the exact dimensions of the defects,enabling them to undertake appropriate welding repairs.In the endeavor to calculate the length,width,and area of these defects,distinguishing different defects within the semantic segmentation image is imperative.For this purpose,a method founded on connected component analysis was utilized in this study.As the nomenclature implies,a connected component encompasses interconnected pixels within an image segment characterized by analogous pixel values.By deploying connected component analysis,individual components within an image can be discerned and marked.This technique has garnered considerable traction across various computer vision tasks.However,to obviate any potential interference caused by fluctuations in pixel values when isolating distinct connected components,images are typically subjected to binarization preceding connected component analysis.Integral to this analysis is the determination of neighborhood connectedness.Two predominant criteria for this assessment are the 4-connected neighborhood and the 8-connected neighborhood,the principles of which are delineated in Fig.12.

    Figure 12:4-connected neighborhood and 8-connected neighborhood of adjacent pixels

    In the realm of the 4-connected neighborhood definition,two pixels can be adjudged as neighboring solely if they exhibit contiguous connections either horizontally or vertically.When pointP0(x,y)is examined under the 4-connected neighborhood premise,the coordinates of its four adjacent pixels emerge asP1(x-1,y),P2(x+1,y),P3(x,y-1),andP4(x,y+1).Conversely,under the 8-connected neighborhood paradigm,besides the horizontal and vertical axes,pixels are also deemed interconnected if they adjoin across the diagonal axes.Accordingly,for the pointP0(x,y),the eight neighboring pixel coordinates,when evaluated under the 8-connected neighborhood definition,areP1(x-1,y),P2(x+1,y),P3(x,y-1),P4(x,y+1),P5(x-1,y-1),P6(x+1,y-1),P7(x-1,y+1),andP8(x+1,y+1).The methodology outlined in this study harnesses the 8-connected neighborhood definition.

    Upon subjecting the rail surface defects to semantic segmentation image analysis using connected components,the OpenCV-Python function cv2.ConnectedComponentsWithStats() is leveraged to discern the connected components.This function yields the count of connected components,the expanse of each individual component,and the dimensions of the encompassing rectangle.The procedural sequence through which defect parameters are derived,rooted in connected component analysis as delineated in this research,is exhibited in Fig.13.

    Figure 13:Connected component analysis process of this article

    Firstly,the semantic segmentation image,in the initial phase,is subjected to denoising via median filtering.Following this,it undergoes transformation into a binary grayscale image.Subsequent to this transformation,a binary image emerges through threshold segmentation.The concluding step embarks upon connected component analysis.From this analysis,several metrics can be derived:the quantity of connected components within each image,pixel markings within the image,and comprehensive statistics corresponding to each marking—encompassing the length,width,and area of the respective contours.To ensure distinctive identification of different connected components,a colormap facilitates the filling of each with a unique hue.Fig.14 exhibits the outcomes of this connected component analysis.It is imperative to note that the defect boundaries used within the connected component analysis are devoid of direction.As such,the included analysis does not delve into boundary directionality.Therefore,the graphics leveraged for connected component analysis in this research are undirected,rendering them neither as strongly nor weakly connected graphs.

    Figure 14:Example of connected component analysis results

    The rail head images featured in this study have an actual width of 73 mm,and all utilized rail surface defect images maintain a dimension of 160 × 370 pixels.Leveraging the proportionality principle,the actual length of the rail within the image is discerned to be approximately 168.81 mm.This translates to a total rail areaSentire_realof roughly 12323.31 mm2.formula(21)outlines the calculation methodology for the real rail defect length,drawing upon the ratio between the defect length in the semantic segmentation imageLdefect_segand the entirety of the image lengthLentire_seg.It’s specified thatLentire_real,representing the real rail length within the image,stands at 168.81 mm.

    According to the percentage of defect area in the semantic segmentation imageSdetect_segto the entire image areaSentire_seg,the actual rail defect areaSdefect_realcan be calculated,as shown in formula(22),whereSentire_realrepresents that the actual area of the rail area corresponding to the semantic segmentation image is 12323.31 mm2.

    Post connected component analysis,the actual dimensions of the defect within the image—its length,width,and area—can be deduced by incorporating the relevant coefficient.Given that the image’s background pixels too form part of a connected component,any component exceeding an area of 30000 pixels is classified as the backdrop.Fig.15 presents the algorithm’s operational outcomes.In this depiction,start_x and start_y signify the x and y coordinates of the starting point of the connected component,respectively.Meanwhile,“width”represents its image width,“height”corresponds to its image length,and “area” delineates its space within the image.The terms “actual width”,“actual height”,and“actual area”convey the real-world dimensions of the connected component.

    Figure 15:Operating results of damage area parameter calculation

    For example,in Fig.14a,an area,depicted in purple and characterized as connected,was found to be 64 pixels in length and 23 pixels in width,resulting in an area of 960 pixels.When translated to real-world dimensions,the defect exhibited an actual length of 29.18 mm,a width of 10.49 mm,and spanned an area of 199.68 mm2.Additionally,another connected region,colored in coffee,measured 131 pixels in length,45 pixels in width,and occupied an area of 3803 pixels.Correspondingly,this defect’s actual dimensions were determined to be 59.74 mm in length,20.52 mm in width,and an area of 791.02 mm2.For Fig.14b,the analysis revealed a connected area measuring 77 pixels in length,55 pixels in width,with a total area of 2899 pixels.The actual measurements of this defect were found to be 35.11 mm in length,25.08 mm in width,and 602.99 mm2in area.In Fig.14c,a blue connected area was identified with a length of 28 pixels,a width of 36 pixels,and an overall area of 650 pixels.In actual measurements,the defect was 12.77 mm long,16.42 mm wide,and spanned an area of 135.20 mm2.A second region,colored in coffee,was 26 pixels in length,41 pixels in width,and covered an area of 819 pixels.The actual measurements of this defect were 11.86 mm in length,18.70 mm in width,and an area of 170.35 mm2.A third region,depicted in purple,was found to be 40 pixels long,28 pixels wide,and had an area of 676 pixels.The real-world measurements of this defect were 18.24 mm in length,12.77 mm in width,and an area of 140.61 mm2.Fig.14d showcased a coffee-colored connected area that was 22 pixels long,21 pixels wide,with an area encompassing 353 pixels.This translated to an actual length of 10.03 mm,a width of 9.58mm,and an area of 73.42 mm2.Another area,colored in purple,had dimensions of 20 pixels in length,22 pixels in width,and covered 361 pixels.The real-world dimensions of this defect were determined to be 9.12 mm in length,10.03 mm in width,and 75.09 mm2in area.

    Through the use of connected component analysis,we are able to determine the actual dimensions of the defect,including its length,width,and area.This provides maintenance personnel with a clear and intuitive understanding of the defect’s parameters,enabling them to undertake targeted maintenance actions.

    6 Conclusion

    In addressing the inefficiencies of conventional machine vision methods and the irregular and intricate defect shapes encountered in rail surface defect detection,an enhanced UPerNet algorithm is presented in this study.Initially,Swin-T is employed as the backbone network to circumvent issues related to inductive bias and the locality of feature extraction inherent in CNN-based backbone networks.Given the influence of BN on the convergence of the semantic segmentation mode,SyncBN is adopted to refine the model,thereby enhancing the training stability during multi-GPU computations.Conclusively,the Lovász-hinge is selected as the loss function to address the challenge posed by cross-entropy loss,which tends to overlook the interdependence between pixels.Experimental findings indicate that the proposed methodology augments detection accuracy across both datasets.Specifically,for the Rail-BY dataset,the PA was recorded at 93.39%,the IoU at 83.69%,the Dice at 91.12%,and the Precision at 90.85%.Similarly,for the RSDDs dataset,the PA,IoU,Dice,and Precision values were determined to be 93.35%,87.58%,93.38%,and 93.41%,respectively,illustrating the method’s proficiency in accurately executing rail surface defect segmentation tasks.

    Subsequent to the semantic segmentation of rail surface defects,connected component analysis technology is utilized to discern distinct defects within a singular image.Thereafter,by enumerating the pixel count,the authentic length and area of each defect zone are deduced.Such insights prove invaluable for railway maintenance personnel in ascertaining the extent of the inflicted damage.

    Acknowledgement:Thanks to our tutors and researchers for their assistance and guidance.The authors also gratefully acknowledge the helpful comments and suggestions of the reviewers,which have improved the presentation.

    Funding Statement:This work is supported in part by the National Natural Science Foundation of China(Grant No.62066024).Gansu Province Higher Education Industry Support Plan(2021CYZC-34),Lanzhou Talent Innovation and Entrepreneurship Project(2021-RC-27,2021-RC-45).

    Author Contributions:Study conception and design: Yongzhi Min,Jiafeng Li,Yaxing Li;data collection:Jiafeng Li;analysis and interpretation of results:Yongzhi Min,Jiafeng Li,Yaxing Li;draft manuscript preparation:Jiafeng Li,Yaxing Li.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产精品野战在线观看| 国产欧美日韩综合在线一区二区| √禁漫天堂资源中文www| 久久国产亚洲av麻豆专区| 黄网站色视频无遮挡免费观看| 97人妻精品一区二区三区麻豆 | 国产av在哪里看| 男女午夜视频在线观看| 亚洲午夜理论影院| 欧美成人性av电影在线观看| videosex国产| 久久中文看片网| 涩涩av久久男人的天堂| 此物有八面人人有两片| 成人特级黄色片久久久久久久| 成人特级黄色片久久久久久久| 亚洲av五月六月丁香网| 亚洲一区中文字幕在线| 美女国产高潮福利片在线看| 少妇 在线观看| 少妇 在线观看| 亚洲国产毛片av蜜桃av| 日本a在线网址| 日本撒尿小便嘘嘘汇集6| 91精品国产国语对白视频| 老司机靠b影院| 一进一出好大好爽视频| 国产精品久久视频播放| 久久国产精品人妻蜜桃| 夜夜看夜夜爽夜夜摸| 少妇被粗大的猛进出69影院| 亚洲五月婷婷丁香| 中文字幕久久专区| 国产一级毛片七仙女欲春2 | 欧美日韩乱码在线| 国产单亲对白刺激| 午夜激情av网站| 一级a爱片免费观看的视频| 亚洲九九香蕉| 一卡2卡三卡四卡精品乱码亚洲| 搞女人的毛片| 国产精品久久久av美女十八| 老鸭窝网址在线观看| 色av中文字幕| 久久国产精品影院| 中文字幕人妻丝袜一区二区| 国产麻豆成人av免费视频| 精品国内亚洲2022精品成人| 男人舔女人的私密视频| 久久久国产精品麻豆| 国产av又大| 国产欧美日韩一区二区精品| 国产男靠女视频免费网站| 色哟哟哟哟哟哟| 男女午夜视频在线观看| 美女高潮喷水抽搐中文字幕| 操出白浆在线播放| 好看av亚洲va欧美ⅴa在| 99国产极品粉嫩在线观看| 90打野战视频偷拍视频| 国产麻豆69| 99久久99久久久精品蜜桃| 黑人巨大精品欧美一区二区蜜桃| 91av网站免费观看| 色av中文字幕| 在线观看www视频免费| 国产精品免费一区二区三区在线| 69精品国产乱码久久久| 中文字幕高清在线视频| 97碰自拍视频| av在线天堂中文字幕| 久久这里只有精品19| 久久久久久久久中文| 波多野结衣高清无吗| 91成年电影在线观看| 亚洲欧美激情综合另类| 手机成人av网站| av欧美777| 亚洲国产精品成人综合色| 国产精品久久电影中文字幕| 精品一区二区三区视频在线观看免费| 美女高潮到喷水免费观看| 女警被强在线播放| 在线十欧美十亚洲十日本专区| 色在线成人网| 一边摸一边抽搐一进一小说| 一级毛片高清免费大全| 亚洲第一av免费看| www.www免费av| 制服人妻中文乱码| 国产精品久久久久久亚洲av鲁大| 男女之事视频高清在线观看| 大码成人一级视频| 亚洲成人免费电影在线观看| or卡值多少钱| www.精华液| av免费在线观看网站| 国产91精品成人一区二区三区| 成人手机av| 亚洲色图综合在线观看| 午夜影院日韩av| 免费女性裸体啪啪无遮挡网站| 中文字幕色久视频| 亚洲 欧美 日韩 在线 免费| 嫩草影院精品99| 九色国产91popny在线| 亚洲三区欧美一区| 91精品三级在线观看| 99在线视频只有这里精品首页| 99精品在免费线老司机午夜| 韩国精品一区二区三区| 俄罗斯特黄特色一大片| 人妻久久中文字幕网| 午夜激情av网站| 岛国视频午夜一区免费看| 别揉我奶头~嗯~啊~动态视频| 久久国产亚洲av麻豆专区| 亚洲av成人不卡在线观看播放网| 午夜亚洲福利在线播放| 久久国产精品人妻蜜桃| 久久久国产欧美日韩av| 久久精品成人免费网站| 丝袜美足系列| 国产免费av片在线观看野外av| 男人的好看免费观看在线视频 | 制服丝袜大香蕉在线| 狂野欧美激情性xxxx| 亚洲色图av天堂| 婷婷六月久久综合丁香| 国产av又大| 久久国产亚洲av麻豆专区| 最好的美女福利视频网| 欧美成人性av电影在线观看| 国产aⅴ精品一区二区三区波| 一区二区三区激情视频| 久久亚洲精品不卡| av视频免费观看在线观看| 免费观看精品视频网站| 50天的宝宝边吃奶边哭怎么回事| 精品卡一卡二卡四卡免费| 欧美成人午夜精品| 国产精品久久久av美女十八| 国产av一区二区精品久久| 久久久久久国产a免费观看| 亚洲欧美一区二区三区黑人| 久久久国产成人免费| 美女扒开内裤让男人捅视频| 丝袜美足系列| 成人av一区二区三区在线看| 在线永久观看黄色视频| 午夜老司机福利片| 亚洲欧美精品综合一区二区三区| 久久婷婷人人爽人人干人人爱 | 亚洲狠狠婷婷综合久久图片| 女生性感内裤真人,穿戴方法视频| 人成视频在线观看免费观看| 国产激情久久老熟女| 波多野结衣巨乳人妻| 欧美 亚洲 国产 日韩一| 久久久国产欧美日韩av| 性色av乱码一区二区三区2| 首页视频小说图片口味搜索| 搞女人的毛片| 国产男靠女视频免费网站| 99国产综合亚洲精品| 男女之事视频高清在线观看| 色综合欧美亚洲国产小说| √禁漫天堂资源中文www| 亚洲精品一卡2卡三卡4卡5卡| 久久国产乱子伦精品免费另类| 亚洲精品粉嫩美女一区| 黑丝袜美女国产一区| 一区二区三区精品91| 性欧美人与动物交配| 窝窝影院91人妻| 久久久久久免费高清国产稀缺| 国产真人三级小视频在线观看| 成人免费观看视频高清| 日本 欧美在线| 在线观看日韩欧美| 黄色a级毛片大全视频| 国产乱人伦免费视频| 免费少妇av软件| 成人亚洲精品一区在线观看| 波多野结衣一区麻豆| 香蕉久久夜色| 国产精品一区二区三区四区久久 | 久9热在线精品视频| 美国免费a级毛片| 两个人免费观看高清视频| 亚洲人成电影免费在线| 老司机在亚洲福利影院| 国产精品av久久久久免费| 国产精品永久免费网站| 男女午夜视频在线观看| 午夜精品久久久久久毛片777| 国产免费男女视频| 久久精品国产清高在天天线| 精品久久蜜臀av无| 国产精品综合久久久久久久免费 | 日日摸夜夜添夜夜添小说| 国产不卡一卡二| 制服诱惑二区| 欧美日韩精品网址| 国产精品自产拍在线观看55亚洲| 午夜免费鲁丝| 他把我摸到了高潮在线观看| 国产97色在线日韩免费| 99久久久亚洲精品蜜臀av| 男女之事视频高清在线观看| 中文字幕高清在线视频| 久久天堂一区二区三区四区| 99久久综合精品五月天人人| 成人国产一区最新在线观看| 大香蕉久久成人网| 成人国产综合亚洲| 免费看美女性在线毛片视频| 精品国内亚洲2022精品成人| 乱人伦中国视频| 中文字幕人成人乱码亚洲影| 91麻豆av在线| 成人亚洲精品一区在线观看| 99国产精品免费福利视频| 成人亚洲精品一区在线观看| av福利片在线| 免费观看精品视频网站| 成人国产综合亚洲| 午夜a级毛片| 91av网站免费观看| 热99re8久久精品国产| 美女扒开内裤让男人捅视频| av免费在线观看网站| 免费av毛片视频| 欧美日本亚洲视频在线播放| 天天添夜夜摸| 亚洲一区高清亚洲精品| 中文字幕最新亚洲高清| 色婷婷久久久亚洲欧美| 此物有八面人人有两片| 欧美日韩亚洲综合一区二区三区_| 亚洲男人的天堂狠狠| 日韩免费av在线播放| 18美女黄网站色大片免费观看| 中文字幕高清在线视频| 亚洲精品美女久久av网站| 欧美中文综合在线视频| 天天躁狠狠躁夜夜躁狠狠躁| 久久这里只有精品19| 亚洲成人免费电影在线观看| 亚洲五月天丁香| 91在线观看av| 日韩三级视频一区二区三区| 9色porny在线观看| 亚洲视频免费观看视频| 欧美另类亚洲清纯唯美| 亚洲欧美日韩无卡精品| 欧美色欧美亚洲另类二区 | 亚洲国产欧美一区二区综合| 亚洲最大成人中文| 少妇裸体淫交视频免费看高清 | 香蕉国产在线看| 少妇熟女aⅴ在线视频| 国产精品九九99| 丰满人妻熟妇乱又伦精品不卡| av天堂在线播放| 中国美女看黄片| 日本三级黄在线观看| 亚洲电影在线观看av| av天堂在线播放| 国产一区二区在线av高清观看| 亚洲人成电影观看| svipshipincom国产片| 欧美日韩精品网址| 午夜a级毛片| 麻豆av在线久日| 成人三级做爰电影| 757午夜福利合集在线观看| av在线播放免费不卡| 久久人妻熟女aⅴ| 色av中文字幕| 久久香蕉精品热| 国产亚洲欧美在线一区二区| 脱女人内裤的视频| 亚洲熟妇中文字幕五十中出| 国产成人啪精品午夜网站| 久久久国产欧美日韩av| 两人在一起打扑克的视频| 老汉色∧v一级毛片| 国产精品久久视频播放| 九色亚洲精品在线播放| 亚洲黑人精品在线| 久久久久国产精品人妻aⅴ院| 国产精品自产拍在线观看55亚洲| 亚洲国产欧美网| 亚洲久久久国产精品| www.熟女人妻精品国产| 亚洲片人在线观看| 老司机福利观看| 国产亚洲精品第一综合不卡| 亚洲人成伊人成综合网2020| 伊人久久大香线蕉亚洲五| 精品欧美国产一区二区三| 黑人巨大精品欧美一区二区mp4| 亚洲av第一区精品v没综合| 亚洲午夜理论影院| 日韩视频一区二区在线观看| 精品国内亚洲2022精品成人| 久久久水蜜桃国产精品网| 丰满人妻熟妇乱又伦精品不卡| 国产又爽黄色视频| 色综合欧美亚洲国产小说| 国产野战对白在线观看| 99久久精品国产亚洲精品| 久久久久久人人人人人| 在线观看免费午夜福利视频| 成人av一区二区三区在线看| 在线观看日韩欧美| 精品电影一区二区在线| 可以在线观看的亚洲视频| 精品一区二区三区视频在线观看免费| 给我免费播放毛片高清在线观看| 亚洲电影在线观看av| 久久精品成人免费网站| 欧美成人午夜精品| 禁无遮挡网站| 亚洲精品国产区一区二| 在线观看午夜福利视频| 久久国产精品影院| 十八禁人妻一区二区| 欧美精品亚洲一区二区| 久久久久久久精品吃奶| 黄片播放在线免费| 国产成人欧美在线观看| 亚洲国产中文字幕在线视频| av免费在线观看网站| 午夜免费观看网址| 国产麻豆成人av免费视频| 久久精品国产综合久久久| 亚洲熟妇熟女久久| 日日爽夜夜爽网站| 亚洲精品粉嫩美女一区| 真人做人爱边吃奶动态| 免费高清在线观看日韩| netflix在线观看网站| 色综合站精品国产| 国产亚洲精品av在线| 亚洲精品粉嫩美女一区| 久久久久久久久中文| 琪琪午夜伦伦电影理论片6080| 国产精品电影一区二区三区| 久久香蕉国产精品| 午夜日韩欧美国产| 亚洲九九香蕉| 香蕉久久夜色| 久久精品国产亚洲av高清一级| 欧美乱色亚洲激情| 亚洲av第一区精品v没综合| 欧美绝顶高潮抽搐喷水| 满18在线观看网站| 亚洲精品一卡2卡三卡4卡5卡| 电影成人av| 老司机在亚洲福利影院| 免费看十八禁软件| 精品国产超薄肉色丝袜足j| 91国产中文字幕| 妹子高潮喷水视频| 亚洲一卡2卡3卡4卡5卡精品中文| 老熟妇仑乱视频hdxx| 欧美人与性动交α欧美精品济南到| 国产午夜福利久久久久久| 国产亚洲av嫩草精品影院| 动漫黄色视频在线观看| 欧美色视频一区免费| 亚洲精品中文字幕在线视频| 久99久视频精品免费| 999久久久国产精品视频| 日本一区二区免费在线视频| 国产男靠女视频免费网站| 亚洲成a人片在线一区二区| 午夜影院日韩av| 国产单亲对白刺激| 日本 av在线| 国产成人啪精品午夜网站| 男女做爰动态图高潮gif福利片 | 不卡一级毛片| 可以在线观看的亚洲视频| 好男人电影高清在线观看| 无人区码免费观看不卡| 亚洲精品一区av在线观看| aaaaa片日本免费| 国产男靠女视频免费网站| 首页视频小说图片口味搜索| 一区福利在线观看| 精品第一国产精品| 国产精品九九99| 在线国产一区二区在线| 精品免费久久久久久久清纯| 国产精品1区2区在线观看.| 中国美女看黄片| 夜夜看夜夜爽夜夜摸| 亚洲色图 男人天堂 中文字幕| 女人被躁到高潮嗷嗷叫费观| 国内精品久久久久久久电影| 波多野结衣av一区二区av| 狂野欧美激情性xxxx| av在线播放免费不卡| 丰满人妻熟妇乱又伦精品不卡| 国产私拍福利视频在线观看| 我的亚洲天堂| 久久精品国产99精品国产亚洲性色 | 精品国产亚洲在线| 可以在线观看的亚洲视频| 国产精品影院久久| 人妻久久中文字幕网| 国产av一区二区精品久久| 国产日韩一区二区三区精品不卡| 女生性感内裤真人,穿戴方法视频| 丝袜美足系列| 亚洲七黄色美女视频| 一边摸一边抽搐一进一出视频| 亚洲黑人精品在线| 亚洲 欧美 日韩 在线 免费| 国产极品粉嫩免费观看在线| 无遮挡黄片免费观看| 欧美另类亚洲清纯唯美| 免费在线观看完整版高清| 啦啦啦韩国在线观看视频| 国产精品爽爽va在线观看网站 | 国产视频一区二区在线看| 大香蕉久久成人网| 欧美+亚洲+日韩+国产| 日本撒尿小便嘘嘘汇集6| 搡老岳熟女国产| 美女扒开内裤让男人捅视频| 亚洲,欧美精品.| 成人欧美大片| 久久中文字幕一级| 国语自产精品视频在线第100页| 桃红色精品国产亚洲av| 自线自在国产av| 少妇的丰满在线观看| 国产主播在线观看一区二区| 夜夜躁狠狠躁天天躁| 亚洲中文av在线| 亚洲国产日韩欧美精品在线观看 | 人妻久久中文字幕网| 久久天堂一区二区三区四区| 中文字幕av电影在线播放| 丝袜美腿诱惑在线| 国产精品99久久99久久久不卡| 嫁个100分男人电影在线观看| 亚洲欧美精品综合一区二区三区| 性色av乱码一区二区三区2| 可以在线观看的亚洲视频| 91麻豆av在线| 国产精品影院久久| 十八禁人妻一区二区| 香蕉久久夜色| 美女高潮到喷水免费观看| 午夜免费鲁丝| 99热只有精品国产| 美女扒开内裤让男人捅视频| 男女做爰动态图高潮gif福利片 | 999久久久精品免费观看国产| 18禁黄网站禁片午夜丰满| 午夜福利18| 国产激情久久老熟女| 深夜精品福利| 亚洲av成人一区二区三| 亚洲黑人精品在线| 黄色 视频免费看| 老司机午夜十八禁免费视频| 国产高清激情床上av| 亚洲欧美日韩高清在线视频| 成人免费观看视频高清| 亚洲午夜精品一区,二区,三区| 国产精华一区二区三区| 此物有八面人人有两片| www.精华液| 亚洲精品美女久久久久99蜜臀| 久久精品国产综合久久久| 一个人观看的视频www高清免费观看 | 男女做爰动态图高潮gif福利片 | 亚洲,欧美精品.| 中文字幕高清在线视频| 亚洲国产精品成人综合色| 欧美日韩黄片免| 精品一区二区三区四区五区乱码| 日本 av在线| 日韩大尺度精品在线看网址 | 黑人巨大精品欧美一区二区mp4| 久99久视频精品免费| 国产在线观看jvid| 黄片播放在线免费| 日韩欧美免费精品| 亚洲人成网站在线播放欧美日韩| 很黄的视频免费| 日韩欧美一区视频在线观看| 免费观看精品视频网站| 亚洲av第一区精品v没综合| 亚洲成人免费电影在线观看| 99久久精品国产亚洲精品| 久久婷婷人人爽人人干人人爱 | 最近最新中文字幕大全免费视频| 国产一级毛片七仙女欲春2 | 欧美国产精品va在线观看不卡| 一区二区三区精品91| av天堂在线播放| 国产精品永久免费网站| 成人亚洲精品一区在线观看| av免费在线观看网站| 国产欧美日韩综合在线一区二区| 男女下面进入的视频免费午夜 | 一本大道久久a久久精品| 亚洲人成电影观看| 欧美日本亚洲视频在线播放| 久久午夜综合久久蜜桃| 丝袜美腿诱惑在线| 亚洲国产高清在线一区二区三 | 成人18禁在线播放| 久久久久久大精品| 国产精品免费视频内射| 国产麻豆69| av视频免费观看在线观看| 麻豆一二三区av精品| 国产成人精品久久二区二区91| 欧美亚洲日本最大视频资源| 99久久综合精品五月天人人| 久久精品影院6| 欧美精品啪啪一区二区三区| 欧美人与性动交α欧美精品济南到| 搡老岳熟女国产| 国内精品久久久久精免费| 国产伦人伦偷精品视频| 亚洲一区二区三区色噜噜| 中文字幕色久视频| 老汉色av国产亚洲站长工具| 亚洲国产精品合色在线| 亚洲人成电影免费在线| 久久亚洲精品不卡| 热re99久久国产66热| 亚洲精华国产精华精| 波多野结衣高清无吗| 亚洲精品国产色婷婷电影| 91成年电影在线观看| 侵犯人妻中文字幕一二三四区| 国产精品秋霞免费鲁丝片| 免费av毛片视频| 韩国av一区二区三区四区| 18美女黄网站色大片免费观看| 神马国产精品三级电影在线观看 | 亚洲欧美激情综合另类| 亚洲成人国产一区在线观看| 欧美乱色亚洲激情| 99在线视频只有这里精品首页| 午夜福利欧美成人| 国产成人免费无遮挡视频| 夜夜爽天天搞| 自线自在国产av| 级片在线观看| 免费观看人在逋| 国产一级毛片七仙女欲春2 | 老司机在亚洲福利影院| 热99re8久久精品国产| 欧美午夜高清在线| 美国免费a级毛片| 如日韩欧美国产精品一区二区三区| 久久国产精品男人的天堂亚洲| 成人欧美大片| 视频在线观看一区二区三区| 亚洲美女黄片视频| 免费观看精品视频网站| 母亲3免费完整高清在线观看| www.自偷自拍.com| 亚洲精品国产色婷婷电影| 香蕉丝袜av| 亚洲成人久久性| 国产精品 欧美亚洲| 亚洲性夜色夜夜综合| 亚洲精品在线美女| 欧美另类亚洲清纯唯美| 91成年电影在线观看| 日本黄色视频三级网站网址| 精品午夜福利视频在线观看一区| 国产精品九九99| 在线观看免费视频日本深夜| 可以免费在线观看a视频的电影网站| АⅤ资源中文在线天堂| 成年版毛片免费区| 美女 人体艺术 gogo| 男女做爰动态图高潮gif福利片 | 老汉色av国产亚洲站长工具| 亚洲色图av天堂| 操美女的视频在线观看| 一二三四社区在线视频社区8| 手机成人av网站| 性少妇av在线| 十八禁网站免费在线| 国产精品99久久99久久久不卡| 夜夜夜夜夜久久久久| 亚洲成av片中文字幕在线观看| 欧美不卡视频在线免费观看 | 久久香蕉激情| 丰满的人妻完整版| 超碰成人久久| 美女 人体艺术 gogo| 国产成年人精品一区二区| 9热在线视频观看99| x7x7x7水蜜桃| www.999成人在线观看| 亚洲精品一卡2卡三卡4卡5卡| 最新在线观看一区二区三区| 91精品三级在线观看| 在线观看66精品国产| 亚洲第一欧美日韩一区二区三区|