• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Robustness benchmark for unsupervised anomaly detection models

    2024-03-23 04:33:26PeiWangWeiZhaiandYangCao
    中國科學技術大學學報 2024年1期

    Pei Wang,Wei Zhai,and Yang Cao,2 ?

    1Department of Automation, University of Science and Technology of China, Hefei 230027, China;

    2Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei 230088, China

    Abstract: Due to the complexity and diversity of production environments,it is essential to understand the robustness of unsupervised anomaly detection models to common corruptions.To explore this issue systematically,we propose a dataset named MVTec-C to evaluate the robustness of unsupervised anomaly detection models.Based on this dataset,we explore the robustness of approaches in five paradigms,namely,reconstruction-based,representation similarity-based,normalizing flow-based,self-supervised representation learning-based,and knowledge distillation-based paradigms.Furthermore,we explore the impact of different modules within two optimal methods on robustness and accuracy.This includes the multi-scale features,the neighborhood size,and the sampling ratio in the PatchCore method,as well as the multi-scale features,the MMF module,the OCE module,and the multi-scale distillation in the Reverse Distillation method.Finally,we propose a feature alignment module (FAM) to reduce the feature drift caused by corruptions and combine PatchCore and the FAM to obtain a model with both high performance and high accuracy.We hope this work will serve as an evaluation method and provide experience in building robust anomaly detection models in the future.

    Keywords: robustness benchmark;anomaly detection;unsupervised learning;automated optical inspection

    1 Introduction

    With the increasing demands on product quality,it is significant to detect surface defects in products during production.In scenarios such as industrial defect detection,there is a high demand for unsupervised anomaly detection models due to the lack of defect samples.Additionally,since the models need to handle image degradation interference during testing,the robustness of the models needs to be evaluated.Many unsupervised anomaly detection methods have been proposed;these methods can be divided into five categories:reconstruction-based[1-3],representation similarity-based[4-6],normalizing flow-based[7,8],self-supervised representation learning-based[9],and knowledge distillation-based[10-12].

    The existing methods can achieve high accuracy on the MVTec[13]dataset that is a relatively simple dataset.However,the robustness of these models is unknown.Due to the temporal and spatial differences between the training and deployment phases,there is no guarantee that the images processed in actual production will have the same quality as the training images and that unpredictable degradation of image quality may occur.For example,devices that run for long periods may wear out,causing the relative position of the camera to the product to deviate from the preset value,which can lead to defocus blur and geometric shifts in the images.Therefore,there is an urgent need for a means of evaluating the robustness of models to ensure production safety.Past studies on the robustness of CNN models have involved image classification[14],object detection[15],semantic segmentation[16],human pose estimation[17],and other domains.Since defect samples are unavailable to unsupervised anomaly detection models during training,specialized research on the robustness of unsupervised anomaly detection models is necessary.

    To address the aforementioned needs,we propose a framework that includes a robust evaluation dataset,evaluation metrics,and a method to improve the robustness of the model.It is worth noting that the dataset and evaluation metrics are also applicable for evaluating the robustness of supervised models.Additionally,based on this framework,we evaluate and analyze the robustness of mainstream unsupervised defect detection models and obtain conclusions about the factors affecting robustness.

    Inspired by Dan et al.[14],we propose a dataset named MVTec-C with different types of corruptions to investigate the robustness of unsupervised anomaly detection methods systematically (see Fig.1).We select the corruption types considering the uncertainties as much as possible in the production scenario,including Gaussian noise,Poisson noise,motion blur,defocus blur,light,contrast,JPEG compression,and geometry,with each corruption corresponding to five severity levels.We consider robustness to mean how well the performance on the original data can be maintained on corrupted data,so we define the relative corruption performance to measure the robustness of the model.

    Fig.1. Samples with different corruption types.The first image is the original image in MVTec.

    To understand the robustness of existing unsupervised anomaly detection models,we select one or two representative methods for evaluation from each of the five classes of anomaly detection methods.We find that the representation similarity-based and knowledge distillation-based methods achieve the best balance between performance and robustness,while the reconstruction-based method shows good robustness at the worst performance.Since PatchCore and Reverse Distillation both have high performance and high robustness,studying of their robustness is valuable for practical applications.Therefore,we investigate the effects of different factors on the robustness of these two methods through exhaustive ablation experiments.We draw several conclusions based on the experimental results: (ⅰ) High-level features can help construct robust feature representations.(ⅱ) Multiscale features are beneficial for improving performance and robustness.(ⅲ) Neighborhood aggregation can not improve the generalization capability of features.(ⅳ) The memory bank can maintain good robustness even with a small sampling ratio,so it is suitable for modeling a normal sample distribution.(ⅴ) Information bottlenecks make the features more compact and thus improve the robustness of reverse distillation.(ⅵ) Multi-scale distillation can consider different defect sizes and hence can improve robustness.

    To construct a robust anomaly detection model,we focus on representation similarity-based methods (which perform best considering robustness and performance simultaneously)and attempt to optimize the feature representation.Defects cause a drift in the features we rely on to discriminate defects.However,corruptions introduce an additional drift that leads to a decrease in the accuracy of detecting defects.Considering the different characteristics of corruptions and defects,we hypothesize that corruptions lead to globally consistent drift,while defects lead to local drift in features.Based on this hypothesis,we propose a feature alignment module (FAM) to reduce the globally consistent drift while preserving the local drift caused by defects,thus obtaining a feature representation that is robust to corruptions.We apply the FAM to Patch-Core and significantly improve its robustness while maintaining high performance.Overall,our contributions are as follows:

    (Ⅰ) We construct a robust benchmark for unsupervised anomaly detection methods,including a dataset with eight corruption types,five severity levels,and metrics to assess robustness.

    (Ⅱ) We evaluate the performance and robustness of mainstream unsupervised anomaly detection methods and find that representation similarity-based and knowledge distillationbased approaches are the best paradigms in terms of performance and robustness.

    (Ⅲ) The different components of the two best-performing methods are studied for ablation,thus helping to understand the impact of different factors on robustness.

    (Ⅳ) We propose a feature alignment module to rectify the corrupted features.Combining the proposed module with PatchCore yields a model with robustness while maintaining high performance.

    2 Related work

    2.1 Unsupervised anomaly detection

    In recent years,researchers’ interest in anomaly detection,especially unsupervised anomaly detection,has increased rapidly,and many datasets and unsupervised anomaly detection algorithms have been proposed.The most frequently used dataset is the MVTec[13]that contains five texture and ten object classes,totaling 5354 images.

    To identify defects from images,unsupervised anomaly detection algorithms should produce either the image-level anomaly score required for anomaly detection or the pixel-level anomaly map as required for the anomaly segmentation or both.The mainstream unsupervised detection algorithms can be classified into five categories: reconstruction-based,representation similarity-based,self-supervised representation learning-based,normalizing flow-based,and knowledge distillation-based.

    Reconstruction.Reconstruction-based approaches typically use generative adversarial networks,autoencoders,and variational autoencoders to reconstruct the input image under the assumption that a model trained on normal samples can correctly reconstruct only normal regions but not defective regions.AnoGAN[1]used generative adversarial networks to learn the distribution of normal images,and the training phase trains a generator of normal image patches.In the test phase,for each patch in the test image,the hidden vector is iteratively adjusted using the discriminant score and the intermediate features of the discriminator as a guide.Finally,the residual map of the test image and the output image of the generator are combined with the residual loss of the discriminator’s intermediate features to identify defects.The f-AnoGAN[18]used an autoencoder to generate the hidden vector,avoiding the problem of time-consuming iterative optimization in AnoGAN.Ganomaly[2]detected defects based on an encoderdecoder-encoder structure by comparing the encoding of the first encoder with the encoding reconstructed by the second encoder.AE-SSIM[3]introduced the structural similarity index measure (SSIM) as the reconstruction loss of the autoencoder.To address the problem that defects are unexpectedly well reconstructed in the autoencoder-based approaches,RIAD[19]proposed training an the autoencoder by recovering images that are partially erased.

    Representation similarity.In representation similaritybased methods,the normal feature distribution is first explicitly modeled during training.K nearest neighbors or Mahalanobis distance is used to calculate the similarity between test features and normal features during testing.Samples with low similarity are considered defective.GaussianAD[20]proposed the use of a multivariate Gaussian model to describe the distribution of pre-trained features of normal samples.SPADE[4]saved patch-level features and image-level features extracted from ResNet;K nearest neighbors searched from image-level normal features are used to compute anomaly scores;and patch features of these selected images are further used to compute patch-level anomaly maps.PaDim[5]estimated a multivariate Gaussian model at each patch position and uses a random dimension selection strategy to reduce the size of the patch features.PatchCore[6]employed a positionindependent feature memory bank and uses a greedy selection strategy to reduce the size of the memory bank.

    Self-supervised representation learning.Several methods[21-23]are dedicated to learning discriminative representations through proxy tasks such as predicting the geometric transformation of the image or contrast learning[24].However,these methods are limited to learning high-level semantic information.CutPaste[9]constructed anomalies by randomly cropping and pasting image blocks and demonstrates that the network’s perception of such anomalies can be generalized to actual defects.

    Normalizing flow.The normalizing flow-based approach uses normalizing flows to transform the normal feature distribution into a Gaussian distribution.The network outputs a probability indicating whether the input is within the normal distribution,and a low probability is expected when the input is defective.DifferNet[7]applied a normalizing flow to imagelevel pre-trained features to obtain the image-level anomaly score.Cflow[8]formed a conditional normalizing flow by incorporating positional encoding in the flow,then the anomaly map is obtained by sliding the conditional normalizing flow over patches.

    Knowledge distillation.In knowledge distillation-based approaches,the teacher model and the student model are expected to produce features with differences in the defective images[10-12,25]that use a set of teacher-student model pairs with different perceptual fields to compute residual maps at multiple scales.Salehi et al.[11]proposed that the distillation of features from an expert network at various layers is preferable to that of a single layer.In these approaches,the teacher and student models use similar structures,leading to a convergence of the teacher and student models’ representations of defects.Deng et al.[12]proposed a reverse distillation paradigm to solve this problem.

    2.2 Evaluating the robustness of anomaly detection models

    In recent years,many works have focused on evaluating the robustness of anomaly-based intrusion detection systems.In intrusion detection scenarios,attacks are applied to anomaly detection models,and the changes in model performance are calculated as robustness criteria.Researchers are dedicated to designing better attacks to evaluate model weaknesses.Goodge et al.[26]designed different types of attacks to evaluate model robustness and improve the robustness to adversarial attacks by optimizing the latent representation.Schneider et al.[27]proposed using different feature robustness metrics,and Han et al.[28]proposed using gray/black-box traffic-space adversarial attacks to evaluate model robustness.To better determine whether the attacked samples are anomalies,Gómez et al.[29]proposed using multiple supporting models and found that the 1D CNN is more robust than LSTM according to their evaluation method.The impact of data poisoning on anomaly detection systems has also been studied[30-32].

    However,these studies on the robustness of anomaly detection have focused mainly on intrusion detection scenarios and are difficult to apply to other scenarios for the following reasons.First,most of these works have evaluated the robustness against adversarial attacks or data pollution,but the robustness against common corruption is also important,especially in industrial defect detection scenarios.Second,anomaly detection models in intrusion detection scenarios mainly address one-dimensional time-series data,while image anomaly detection models process multi-dimensional image data,and robust models need to adapt to more complex environmental changes.To the best of our knowledge,research on the robustness of image-based anomaly detection in scenarios such as industrial defect detection is lacking.

    2.3 Benchmarking robustness to common corruptions

    Many methods have been proposed for studying the robustness of CNNs against common image corruptions[14-17,33,34].To investigate the robustness of different classification networks against common image corruptions and perturbations,Dan et al.[14]proposed two datasets,ImageNet-C and ImageNet-P.Michaelis et al.[15]provide a benchmark for evaluating the performance of a target detection model in the face of image corruptions.Similarly,Kamann et al.[16]accessed the robustness of the semantic segmentation model to real-world image corruptions.Altindis et al.[33]evaluated instance segmentation models with real-world image corruptions and out-of-domain images.Wang et al.[17]constructed three robustness benchmarks to study the drawbacks of human pose estimation models.

    3 Methods

    3.1 Robustness benchmark

    We construct the robust benchmark dataset MVTec-C by adding eight corruption types with five severity levels to each image in the MVTec test set.The corruption types include Gaussian noise,Poisson noise,motion blur,defocus blur,light,contrast,JPEG compression,and geometry.Fig.1 shows some samples from the MVTec-C.

    The performance of an anomaly detection model is usually measured using the area under the receiver operator characteristic (AUC).Specifically,image-level AUC is used for imagelevel anomaly detection,while the performance of anomaly segmentation is determined by the pixel-level AUC.Referring to Michaelis et al.[15],the performance on corrupted data is measured by the mean performance under corruption(mPC):

    Here,Ncdenotes the total number of different corruption types included in the test,andNsrepresents the number of severity levels for each type of corruption.In this context,cindexes the specific type of corruption,andsindicates the severity level of the corruption.The robustness of a model is defined as its ability to maintain original performance levels under various corruption scenarios.To quantify this,we introduce the relative performance under corruption (rPC),defined as:

    In the above equation,AUCcleanrefers to the model’s AUC performance on clean,uncorrupted data.Therefore,rPC provides a comparative measure of the model’s robustness by evaluating its performance under corrupted conditions relative to its performance on clean data.

    3.2 Evaluation of representative methods

    We evaluate several of the most representative contemporary unsupervised anomaly detection methods using our benchmark and discuss the advantages and disadvantages of these methods.Then,we select two methods that perform best in terms of both performance and robustness combined and investigate the impact of different components of these methods on the models’ robustness.

    3.2.1 Comparison between different methods

    The f-AnoGAN[18]used an encoder to map the image patches into the hidden space and then uses the WGAN to recover the image patches.The output image of the generator and the intermediate features of the discriminator are used to compute the anomaly score.PaDim[5]used features from multiple semantic layers in the ImageNet pre-trained model,described using a multivariate Gaussian model at each patch position,and used the Mahalanobis distance to compute the anomaly score at each location to generate the anomaly map.PatchCore[6],based on ImageNet pre-trained features,used a greedy strategy to keep the maximal representative set of normal patch features and uses kNN to compute the anomaly score for each test patch.CutPaste[9],a data augmentation method is used in CutPaste to learn self-supervised features,a multivariate Gaussian model is used to model the distribution of global features,and the image-level anomaly score is computed using the Mahalanobis distance.Cflow[8]uses conditional normalizing flow to explicitly estimate the anomaly score at each patch location.Reverse Distillation[12]uses teacher-student models to learn the distribution of normal features at multiple scales.The feature maps of the teacher and student models are compared to obtain the anomaly map.

    3.2.2 In-depth study of specific methods

    To better investigate robustness,it is essential not only to evaluate the mainstream methods’ performance in robustness but also to understand the factors that affect robustness and how they influence it.Ablation experiments can be conducted to investigate these factors and achieve this goal.However,the significant differences among the different methods make it difficult to study these factors in a unified framework.Our benchmark testing revealed that PatchCore by Roth et al.[6]and Reverse Distillation by Deng et al.[12]outperform other methods in both accuracy and robustness and thus warrant further research.Therefore,we conducted more thorough research on these two methods separately to explore the factors that affect robustness.

    PatchCore.The PatchCore method workflow is shown in Fig.2.First,local patch features are extracted and aggregated into a memory bank.The mid-level patch features are utilized to minimize bias toward ImageNet classes,while a feature aggregation over a local neighborhood ensures retention of sufficient spatial context.Second,the coreset-reduction method is introduced to reduce redundancy in the extracted patch-level memory bank and significantly reduce the storage memory and inference time.This is achieved through greedy coreset subsampling for nominal feature banks.Finally,the nearest neighbors of the test features are retrieved from the memory bank and compared with the test patch features to compute anomaly scores.We explore the robustness of the model from the following aspects:

    Fig.2. The pipeline of PatchCore[6].Neighborhood aggregation enhances the patch representations extracted from a pre-trained model.Then the greedy sampling constructs a compact yet maximal representative memory bank.Finally,each patch is scored by calculating the distance between the patch representation and its nearest neighbor in the memory bank.

    ● We explore the impact of features of different scales as patch representations.

    ● Different neighborhood sizes in neighborhood aggregation are compared.

    ● We compare the results when sampling with different sampling ratios to obtain the memory bank.

    Reverse Distillation.Fig.3 shows the flow of Reverse Distillation.First,a pre-trained deep neural network is used as a teacher model to extract the feature representations of the original input data.Second,multi-scale feature fusion (MFF)and a method called “one-class embedding” (OCE) are used to transform the extracted features into low-dimensional,compact representations.Third,a student model is used as a decoder to reconstruct the teacher model’s representations at different scales from the low-dimensional feature embeddings.Fourth,the output of the student model is compared with the output of the teacher model,and the anomaly score and anomaly localization information are obtained by calculating the reconstruction error.We explore several aspects of the factors affecting the robustness of Reverse Distillation:

    Fig.3. The pipeline of Reverse Distillation[12].First,the teacher model acts as a feature extractor and outputs multi-scale features.Secondly,multi-scale features are merged and compressed into low-dimensional embeddings by the multi-scale feature fusion (MFF) and the one-class embedding (OCE) module.Third,the student model tries to reconstruct the multi-scale features output by the teacher model.Finally,the cosine distance is used to compare the outputs of the student and teacher models,resulting in an anomaly map.

    ● Different scales of features are aggregated and compared in the MFF module.

    ● We compare the base model,model with OCE,and models with both MFF and OCE.

    ● Features at different scales are distilled and compared.

    3.3 Robust anomaly detection model

    To construct a robust anomaly detection model,we try to optimize the feature representation of the representation similarity-based approach,which yields the best performance.Corruptions cause features to deviate from the original distribution,thus making the use of the original discriminant function to distinguish defective samples from normal samples less effective.We assume that the feature drift caused by corruption is globally consistent,while the feature drift caused by defects is local.This assumption is motivated by the fact that,unlike natural scenes,in fields such as industrial defect detection and medical imaging,it is often necessary to analyze and process images of small regions to detect defects or lesions.It is reasonable to assume that the corruptioninduced feature drift is globally consistent in such small regions.In contrast,defects or lesions typically results in mutations in a smaller local area of the image,which introduces uneven feature drift.Therefore,the average drift of all patch features in an image can be used to estimate the former but not the latter.By minimizing the average drift,we can compensate significantly for corruption-induced drift.Although,this approach will alter defect features,it is not enough to compensate for the defect drift.Thus,a feature alignment module can be designed to recover uncorrupted features by minimizing the average drift.We illustrate the concept in Fig.4.To compute the average drift,we need to obtain the features before degradation.Since obtaining such features is not feasible,we approximate uncorrupted features by selecting K nearest uncorrupted neighbors from the memory bank.

    Fig.4. Corruptions cause a globally consistent drift,while defects cause a local drift from the normal distribution.The average shift of all image patch features tends to align more closely with the drift caused by corruption than that caused by defects.Consequently,minimizing the average shift can effectively reduce the drift induced by corruption by aligning the corrupted features with the normal ones.Although this approach can alter the defect features,it remains insufficient to compensate for the drift caused by defects.

    Let the corruption transformation beG,the inverse transformation beH,the normal patch feature bex,and the degraded patch feature beThen,we havex=H(G(x))=H.If the inverse transformationHcan be estimated,the featurexcan be returned.We introduce the FAM in the test phase as an estimate of the inverse transformationH;then,the original feature can be estimated asx?=FAMM,whereMrepresents for the parameters of the FAM.We optimize the FAM to minimize the distance between patch features and their reference features,as follows :

    whereXrefers to all the patch features in a single image,andXrefrepresents the reference features of the test features.Each test feature corresponds tokreference features selected from the memory bank with kNN.The Euclidean distance between the features transformed by the FAM and their nearest neighbor features in the memory bank is computed to obtain the anomaly score.

    The parameterMof the FAM consists of the coefficient vectorWand the bias vectorB,defined as FAM(x)=WTx+B.It is designed to fit simple transformations,thus avoiding learning semantic information.In addition,the number of iterationsnof the optimization process needs to be limited to prevent overfitting.In summary,our FAM module is

    designed with parametersWandBthat need to be optimized online for each test sample during the testing phase,and two other hyperparameters,including the number of reference featureskand the number of iterationsn.

    4 Experiments

    4.1 Implementation details

    For robustness evaluation,we use publicly available implementations of these methods. Except for f-AnoGAN,WideResnet50 is selected as the backbone network by default.The input image size is 256×256.For PaDim,550 dimensions are randomly selected from features generated by the backbone.For PatchCore,the memory bank constructed at a sampling ratio of 0.01,and the neighborhood size was 3.For CutPaste,we used a 3-way augmentation method with a two-layer MLP network for the projection head.For Cflow,three pooling layers and eight coupling layers are used.For Reverse Distillation,when fusing features of different scales,the number of output channels of OCE is not mentioned in the original paper,and we set it to a fixed value of 2048.For f-AnoGAN,the dimensionality of the hidden vector is set to 100.When combining the proposed FAM with PatchCore,the FAM transforms the extracted features and calculates the anomaly score by performing the kNN on the memory bank.The initial parameters of the FAM areW=1andB=0.Thus it is equivalent to the original PatchCore before FAM is optimized.We use Adam optimizer to adjust the parameters of FAM and set the learning rate to 0.01.All experiments are performed on four NVIDIA GeForce RTX 3090 GPUs.

    4.2 Robustness study

    4.2.1 Comparison of different methods

    The comparison of the results of different methods is shown in Table 1.Based on the experimental results,we make the following analysis: (ⅰ) f-AnoGAN[18]has the lowest performance but the highest robustness.We speculate that this is not because the decoder reconstructs the degraded image well but because,in the reconstruction-based paradigm,degradation interferes little with the identification of “easy” defects.(ⅱ) Although both PaDim[5]and PatchCore[6]are methods based on representation similarity,their robustness differs significantly.It is hypothesized that the distribution of the location-dependent patch features of the training samples does not generalize well to the new data due to the lack of location diversity of the images in MVTec.In contrast,a uniform location-independent memory bank is more generalizable.Also,according to Ref.[8],the multivariate Gaussian model does not fit the complex distribution well.(ⅲ) CutPaste[9]also adopts a multivariate Gaussian model,but its image-level robustness greatly exceeds PaDim.This indicates that the features obtained by CutPaste through self-supervised learning are more focused on semantic information and,therefore,less subject to interference from non-semantic degradation.(ⅳ) As one of the highest-performance methods on clean data,Cflow[8]does not perform well in terms of robustness.We speculate that it is because the conditional normalizing flow is sensitive to feature degradation.(ⅴ) Reverse Distillation[12]and PatchCore achieve the best balance between accuracy and robustness.To understand the robustness of these two methods in more detail,we perform more in-depth experiments and analyses below.

    Table 1. Comparison of the results of the different methods.“clean” refers to the accuracy (AUC) on MVTec.The image-level and pixel-level metrics are displayed for each method.For some methods (f-AnoGAN,CutPaste),we do not find implementations provided by the authors,and the publicly available implementations we used only generate image-level results.

    4.2.2 In-depth exploration of PatchCore

    We analyze the results of multi-scale features,sampling ratio,and neighborhood size,respectively.

    Multi-scale features.The results of using features of different scales are shown in Fig.5a.Layer2 and layer3 features have significantly better robustness than layer1.It is reasonable to assume that features from higher semantic levels are less disturbed by the degradation of image quality and thus have better robustness.As for detection rPC,layer2 is slightly worse than layer3 because layer2 is better at detecting defects,which in turn improves its robustness.

    Fig.5. Analysis experiments.(a) The effect of multi-scale features in PatchCore on robustness.“1”,“2”,and “3” represent the features from layer1,layer2,and layer3 in ResNet,respectively.(b) The effect of neighborhood size on the robustness of the domain aggregation module in PatchCore.(c) The effect of multi-scale features on robustness in Reverse Distillation.(d) Effect of multi-scale distillation in Reverse Distillation on robustness.(e) Effect of the number of FAM iterations.(f) Effect of the number of FAM reference features.

    Sampling ratio.The effect of the sampling ratio is shown in Table 2.The robustness increases with the sampling ratio,indicating that increasing the number of sampled features helps to generalize the memory bank to degenerate features.Even if only 0.01 of normal features are sampled,the robustness is only slightly lower than when 0.4 of normal features are sampled,indicating that the memory bank obtained by greedy sampling is a good representation of the original normal sample and thus a robust way to model the original normal distribution.

    Table 2. Impact of samplings ratio in PatchCore.

    Neighborhood size.As shown in Fig.5b,when the neighborhood size exceeds 2,the robustness gradually decreases as the neighborhood size increases.Though neighborhood aggregation improves the performance[6],it does not help improve robustness.A reasonable speculation is that aggregating more local information does not alleviate the feature shift caused by corruptions.At the same time,too large neighborhood size leads to patch features being damaged,which is detrimental to anomaly detection.

    4.2.3 In-depth exploration of Reverse Distillation

    The results of multi-scale features,OCE and MFF module,and multi-scale distillation are analyzed.

    Multi-scale features.As shown in Fig.5c,using higher semantic features leads to better robustness,similar to PatchCore[6].This is because the higher-level features are less disturbed by corruptions and closer to the original feature distribution,thus maintaining better robustness,which also explains the better performance of using features from layer3 than multi-scale features.

    OCE and MFF module.As shown in Fig.6,the OCE module improves the robustness,but the MFF module does not contribute to the robustness.Because OCE provides an information “bottleneck” so that only the critical information needed to reconstruct the teacher’s network response is retained,thus filtering out non-critical low-level information about degradation.Additionally,the robustness is slightly reduced by introducing the MFF since MFF introduces lowlevel features that degrade more heavily.

    Fig.6. Impact of Reverse Distillation OCE module and MFF module on robustness.“Plain” represents the base model.

    Multi-scale distillation.The impact of the distillation of features at different scales on the robustness is shown in Fig.5d,and it is worth noting that layer1 > layer3 >layer2 in terms of robustness,and it is necessary to analyze the mechanism.Due to its reverse distillation paradigm,the highresolution features of the student model do not suffer from more degradation interference than the lower-level features.In contrast,the high-resolution features of the teacher model suffer more degradation interference than the low-resolution features.However,the high-resolution features have more detailed information and thus facilitate the segmentation of defects.Therefore,layer1 and layer3 have better robustness than layer2,and aggregating features of all scales can achieve higher robustness.

    4.3 Robust anomaly detection method

    We combine the proposed FAM with PatchCore[6]to obtain a new model,and the evaluation results are shown in Table 3,wherek=2 andn=8.The results show that the FAM improves robustness for most corruptions types.For example,compared to PatchCore,the image-level rPC and pixellevel rPC under the JPEG compression category are improved by 5.9% and 3.9%,respectively.In addition,there is no significant improvement in robustness under the geometry category since the geometric transformation does not result in a globally consistent feature drift,which confirms that FAM mainly serves to reduce the globally consistent feature drift.Although the robustness of the model has been significantly improved,there are still limitations.First,the pixel-level AUC on the clean dataset (MVTec) decreased by 0.4%.This is because,in the absence of defects,the iteration of the FAM causes the defect features to overfit normal features.However,the model still maintains high accuracy,and the robustness has been significantly improved,so this compromise is acceptable.Second,the image-level rPC in the geometric category decreased by 1.1%.This is because the simple structure of FAM cannot capture the complex transformations caused by image rotation.Specifically,CNN lacks rotation invariance.When the image rotates,the image pixels that affect a patch feature and the combination of pixels and convolution kernel parameters change.Hence,the resulting changes in patch features are complex,exceeding the learning ability of FAM’s linear structure.Increasing the complexity of FAM and introducing more contextual information for patch features may help solve this problem.Fig.7 shows the comparisons between PatchCore and our method on some samples.

    Table 3. Comparison of the results of different methods.“clean” refers to the performance on MVTec.The image-level results and pixel-level results are shown for each method.The last column shows the improvement of our method.Blue indicates a decrease,green indicates an improvement of less than 5%,and red indicates a significant improvement of more than 5%.

    Fig.7. Qualitative comparisons between PatchCore and our method.

    4.3.1 Effect of the number of reference features

    The robustness of different reference feature numbers is compared to Fig.5f.k=2 is slightly better thank=1,and the robustness does not change significantly whenk>2,which indicates that when the sampling ratio of 0.01,the memory bank still maintains good densities to represent the normal distribution.We speculate that a larger number of reference features will help improve the robustness when the sampling ratio is further reduced.

    4.3.2 Effect of the number of iterations

    As shown in Fig.5e,when the number of iterations is small,the robustness increases subsequently,but further increasing the number of iterations leads to a decrease in robustness.A plausible reason for this is overfitting.FAM uses a simple model to fit the mapping between degenerate feature distributions and normal feature distributions.It is difficult to recover the original features,so we use the nearest neighbors from the normal memory bank instead.However,there aredifferences between the nearest neighbors and the original features,so too many iterations will damage the feature representation and thus degrade the model’s performance.Thereforen=8 may be a good compromise.

    5 Conclusions

    In this paper,a comprehensive benchmark is constructed to evaluate the robustness of unsupervised anomaly detection models based on real-world corruptions in production scenarios.Thanks to this benchmark,we find many meaningful conclusions and a way to improve robustness.First,the mainstream unsupervised anomaly detection methods in five different paradigms are evaluated,and the advantages and disadvantages of each paradigm are analyzed.Then,the two approaches that perform best in terms of performance and robustness are investigated through careful ablation experiments,and the effects of different modules on the robustness of these two models are explored.Finally,a module is proposed to eliminate the feature drift caused by corruptions,and its effectiveness in improving robustness is demonstrated experimentally.

    Acknowledgements

    This work was supported by National Natural Science Foundation of China (62306295).

    Conflict of interest

    The authors declare that they have no conflict of interest.

    Biographies

    Pei Wangis currently working at Alibaba Cloud in Hangzhou.He received his master’s degree in Control Science and Engineering from the University of Science and Technology of China in 2023.His research interests focus on computer vision.

    Yang Caois currently an Associate Professor in the Automation Department at the University of Science and Technology of China.He received his Ph.D.degree in Pattern Recognition and Intelligent System from the Northeastern University in 2004.His research interests include computer vision and multimedia processing.

    国产欧美日韩精品一区二区| 毛片一级片免费看久久久久| 日韩av在线免费看完整版不卡| 久久毛片免费看一区二区三区| 国产精品蜜桃在线观看| 一级黄片播放器| 我要看黄色一级片免费的| 在线观看美女被高潮喷水网站| 这个男人来自地球电影免费观看 | 国产成人精品无人区| 久久国产精品男人的天堂亚洲 | 在线播放无遮挡| 嘟嘟电影网在线观看| 美女内射精品一级片tv| av在线播放精品| 午夜久久久在线观看| 这个男人来自地球电影免费观看 | 美女主播在线视频| 欧美三级亚洲精品| 国产成人午夜福利电影在线观看| 国产精品久久久久久久久免| 久久午夜福利片| 18禁裸乳无遮挡动漫免费视频| 欧美xxⅹ黑人| 久久av网站| 精品亚洲成a人片在线观看| 一级黄片播放器| 成人无遮挡网站| 亚洲精品国产成人久久av| 搡女人真爽免费视频火全软件| 精品熟女少妇av免费看| 亚洲性久久影院| 高清在线视频一区二区三区| 国产精品99久久99久久久不卡 | 18+在线观看网站| 久久精品国产鲁丝片午夜精品| 狂野欧美白嫩少妇大欣赏| 久久午夜综合久久蜜桃| 视频区图区小说| 交换朋友夫妻互换小说| 久久亚洲国产成人精品v| 日日啪夜夜爽| 国产无遮挡羞羞视频在线观看| 亚洲欧美精品专区久久| 亚洲天堂av无毛| 免费高清在线观看视频在线观看| 国产真实伦视频高清在线观看| 欧美精品一区二区免费开放| 黄片无遮挡物在线观看| 美女主播在线视频| 日本午夜av视频| 久久久国产精品麻豆| 国产午夜精品一二区理论片| 精品人妻一区二区三区麻豆| 色视频www国产| av不卡在线播放| 视频中文字幕在线观看| 色吧在线观看| 91aial.com中文字幕在线观看| 交换朋友夫妻互换小说| 国产精品.久久久| 少妇被粗大的猛进出69影院 | 久久久a久久爽久久v久久| 天堂俺去俺来也www色官网| 好男人视频免费观看在线| 一级毛片电影观看| 色吧在线观看| 欧美日韩精品成人综合77777| 国产成人精品久久久久久| 十八禁网站网址无遮挡 | 成人国产av品久久久| 亚洲成人av在线免费| 国产伦理片在线播放av一区| 男人添女人高潮全过程视频| 精品人妻一区二区三区麻豆| 免费看av在线观看网站| 国产 精品1| 国产淫片久久久久久久久| 国产成人免费无遮挡视频| 色哟哟·www| 久久久久视频综合| 99久久精品热视频| 亚洲精品,欧美精品| 国产精品一区二区性色av| av有码第一页| 日日摸夜夜添夜夜添av毛片| 精品亚洲乱码少妇综合久久| 国产精品久久久久久久久免| 亚洲精品日本国产第一区| 99九九在线精品视频 | 男人添女人高潮全过程视频| 高清在线视频一区二区三区| 久久精品久久久久久久性| 老司机影院成人| 高清欧美精品videossex| 一级毛片久久久久久久久女| 精品久久久久久久久av| 97超视频在线观看视频| 精品久久久久久电影网| 午夜av观看不卡| av在线观看视频网站免费| 少妇高潮的动态图| 伦精品一区二区三区| 熟女av电影| av视频免费观看在线观看| 精品亚洲成a人片在线观看| 亚洲国产成人一精品久久久| 在线观看www视频免费| a 毛片基地| 水蜜桃什么品种好| 日本欧美国产在线视频| 亚洲,一卡二卡三卡| 免费大片18禁| 18禁在线播放成人免费| 久久久久久久精品精品| 亚洲精品国产色婷婷电影| 777米奇影视久久| 亚洲国产av新网站| 视频区图区小说| 国产真实伦视频高清在线观看| 亚洲av国产av综合av卡| 天天躁夜夜躁狠狠久久av| 插逼视频在线观看| 免费观看av网站的网址| 亚洲成人av在线免费| 久久久久久久久久久免费av| 国产在线免费精品| 三级经典国产精品| .国产精品久久| 亚洲国产精品成人久久小说| 日韩 亚洲 欧美在线| 丝瓜视频免费看黄片| 久久久精品免费免费高清| 国产精品一区二区三区四区免费观看| 国产免费福利视频在线观看| 欧美xxⅹ黑人| 亚洲成色77777| 亚洲第一区二区三区不卡| 欧美亚洲 丝袜 人妻 在线| 人人妻人人澡人人看| 又粗又硬又长又爽又黄的视频| 男女边吃奶边做爰视频| 噜噜噜噜噜久久久久久91| av在线播放精品| 成人18禁高潮啪啪吃奶动态图 | 国产精品不卡视频一区二区| 新久久久久国产一级毛片| 日日啪夜夜爽| 亚洲美女搞黄在线观看| 亚洲av综合色区一区| 久久久久久久国产电影| 高清在线视频一区二区三区| 免费黄频网站在线观看国产| 国产在线一区二区三区精| av黄色大香蕉| 国产一区有黄有色的免费视频| 观看免费一级毛片| 亚洲第一av免费看| 精品久久久噜噜| 另类亚洲欧美激情| 国产欧美日韩精品一区二区| 久久久精品免费免费高清| 日韩欧美一区视频在线观看 | 天天躁夜夜躁狠狠久久av| 国产成人精品无人区| 22中文网久久字幕| 日韩一本色道免费dvd| 国内少妇人妻偷人精品xxx网站| 这个男人来自地球电影免费观看 | 亚洲精品中文字幕在线视频 | 人人妻人人爽人人添夜夜欢视频 | 亚洲熟女精品中文字幕| 日韩av不卡免费在线播放| 日韩人妻高清精品专区| 日韩一本色道免费dvd| 欧美日韩精品成人综合77777| 亚洲av成人精品一二三区| 黑人高潮一二区| 少妇的逼好多水| 青青草视频在线视频观看| 中文字幕av电影在线播放| 春色校园在线视频观看| 精品卡一卡二卡四卡免费| 国产成人a∨麻豆精品| 性色avwww在线观看| 夜夜骑夜夜射夜夜干| 国产成人aa在线观看| 欧美xxxx性猛交bbbb| 在线观看www视频免费| 成年av动漫网址| 丁香六月天网| 人妻一区二区av| 免费观看av网站的网址| 亚洲精品乱码久久久久久按摩| 2022亚洲国产成人精品| 欧美日韩精品成人综合77777| 久久精品久久久久久噜噜老黄| 国产欧美另类精品又又久久亚洲欧美| 亚洲国产欧美在线一区| 色哟哟·www| 国产高清国产精品国产三级| 国产精品久久久久久av不卡| av女优亚洲男人天堂| 香蕉精品网在线| av线在线观看网站| 久久狼人影院| 人妻一区二区av| 五月开心婷婷网| 午夜福利网站1000一区二区三区| 一级,二级,三级黄色视频| 哪个播放器可以免费观看大片| 搡女人真爽免费视频火全软件| 亚州av有码| 9色porny在线观看| 久久国产亚洲av麻豆专区| 性色avwww在线观看| 中文乱码字字幕精品一区二区三区| 国产中年淑女户外野战色| 中文字幕亚洲精品专区| 乱码一卡2卡4卡精品| 人妻 亚洲 视频| 国产成人freesex在线| 在线看a的网站| 最新的欧美精品一区二区| 国产成人91sexporn| 中文精品一卡2卡3卡4更新| 噜噜噜噜噜久久久久久91| 最近中文字幕高清免费大全6| 中文字幕人妻丝袜制服| 免费播放大片免费观看视频在线观看| 亚洲,一卡二卡三卡| 免费观看的影片在线观看| 免费黄色在线免费观看| 亚洲精品国产av蜜桃| 精品亚洲成a人片在线观看| 欧美老熟妇乱子伦牲交| 熟女电影av网| 久久99蜜桃精品久久| 少妇人妻久久综合中文| 色哟哟·www| 精品少妇内射三级| 亚洲精品成人av观看孕妇| 99热这里只有精品一区| 80岁老熟妇乱子伦牲交| 岛国毛片在线播放| 在线看a的网站| 婷婷色综合大香蕉| 精品午夜福利在线看| 亚洲色图综合在线观看| 一本一本综合久久| 一级av片app| 在线观看免费视频网站a站| h日本视频在线播放| 国产精品一区二区性色av| 国产极品粉嫩免费观看在线 | 亚洲成色77777| 久久久亚洲精品成人影院| 亚洲熟女精品中文字幕| 国产精品欧美亚洲77777| 国模一区二区三区四区视频| 最近的中文字幕免费完整| 丝瓜视频免费看黄片| 久久亚洲国产成人精品v| 美女脱内裤让男人舔精品视频| 欧美精品一区二区免费开放| 久久人人爽人人爽人人片va| 久久99热这里只频精品6学生| 久久久久精品性色| 免费黄网站久久成人精品| 久久精品熟女亚洲av麻豆精品| 色94色欧美一区二区| 国产 一区精品| 国产男人的电影天堂91| 亚洲情色 制服丝袜| 亚洲精品国产av成人精品| 青春草国产在线视频| 亚洲第一av免费看| 色94色欧美一区二区| 日本午夜av视频| 九色成人免费人妻av| 亚洲婷婷狠狠爱综合网| 纯流量卡能插随身wifi吗| 欧美老熟妇乱子伦牲交| 在线观看av片永久免费下载| 日日撸夜夜添| 高清不卡的av网站| 亚洲欧美一区二区三区国产| av黄色大香蕉| 免费大片黄手机在线观看| 国产精品无大码| 有码 亚洲区| 婷婷色综合www| 妹子高潮喷水视频| 日韩精品免费视频一区二区三区 | 美女国产视频在线观看| 免费观看性生交大片5| 亚洲欧美日韩另类电影网站| 丝袜喷水一区| av天堂中文字幕网| 亚洲欧美精品专区久久| 91久久精品国产一区二区三区| a 毛片基地| 另类亚洲欧美激情| 我的老师免费观看完整版| 亚洲怡红院男人天堂| 综合色丁香网| 久久人人爽人人片av| av专区在线播放| 久久免费观看电影| 少妇的逼好多水| 久久国内精品自在自线图片| 蜜桃久久精品国产亚洲av| 免费大片黄手机在线观看| 国产一区二区三区av在线| 夫妻午夜视频| 久久久久网色| 丝瓜视频免费看黄片| 亚洲久久久国产精品| 搡女人真爽免费视频火全软件| 日韩免费高清中文字幕av| 青春草视频在线免费观看| 九九在线视频观看精品| 一边亲一边摸免费视频| 中文乱码字字幕精品一区二区三区| 又爽又黄a免费视频| 国产白丝娇喘喷水9色精品| 亚洲欧美成人综合另类久久久| 亚洲,一卡二卡三卡| 色视频www国产| 国产成人一区二区在线| 深夜a级毛片| 一级毛片久久久久久久久女| 成人二区视频| 亚洲性久久影院| av福利片在线| 六月丁香七月| 在线观看免费视频网站a站| 亚洲熟女精品中文字幕| 最近中文字幕高清免费大全6| 亚洲精品乱码久久久v下载方式| 成年美女黄网站色视频大全免费 | 亚州av有码| 高清欧美精品videossex| 久久久国产欧美日韩av| 亚洲电影在线观看av| 女人精品久久久久毛片| 一本大道久久a久久精品| 丝瓜视频免费看黄片| 少妇丰满av| 亚洲va在线va天堂va国产| 99热这里只有精品一区| 亚洲久久久国产精品| 91aial.com中文字幕在线观看| 国产精品免费大片| 男人爽女人下面视频在线观看| 免费少妇av软件| 交换朋友夫妻互换小说| 久久久国产一区二区| 国产精品久久久久久精品电影小说| 精品一品国产午夜福利视频| 亚洲精品亚洲一区二区| 中国国产av一级| 国产精品久久久久久精品古装| 欧美人与善性xxx| 国产一区二区在线观看日韩| 亚洲国产最新在线播放| 久久精品国产亚洲av天美| 全区人妻精品视频| 日韩大片免费观看网站| 插阴视频在线观看视频| 精品亚洲成a人片在线观看| 波野结衣二区三区在线| 搡女人真爽免费视频火全软件| 99久久综合免费| 日本91视频免费播放| 国内少妇人妻偷人精品xxx网站| 成人特级av手机在线观看| 日本黄色片子视频| 欧美日本中文国产一区发布| 最新中文字幕久久久久| 91精品伊人久久大香线蕉| 亚洲精品一区蜜桃| 黑丝袜美女国产一区| 欧美精品国产亚洲| 国产精品久久久久久精品电影小说| 日本91视频免费播放| 国产免费福利视频在线观看| 又粗又硬又长又爽又黄的视频| 热99国产精品久久久久久7| 久久国产精品男人的天堂亚洲 | 免费观看性生交大片5| 午夜精品国产一区二区电影| 国产黄片美女视频| 亚洲欧洲国产日韩| 午夜老司机福利剧场| tube8黄色片| 日韩一本色道免费dvd| 久久久久人妻精品一区果冻| 亚洲av综合色区一区| 女性生殖器流出的白浆| 亚洲精品乱久久久久久| 久久精品夜色国产| 精品少妇黑人巨大在线播放| 国产一区亚洲一区在线观看| 婷婷色av中文字幕| 免费观看在线日韩| 综合色丁香网| 久久久精品免费免费高清| 免费观看a级毛片全部| 一区二区三区乱码不卡18| 国产成人一区二区在线| 又大又黄又爽视频免费| 国产精品一区www在线观看| 国产精品一区二区在线观看99| 一级二级三级毛片免费看| 久久精品熟女亚洲av麻豆精品| 女人精品久久久久毛片| 成人二区视频| av.在线天堂| 日韩强制内射视频| 高清在线视频一区二区三区| 欧美精品一区二区大全| 日本av手机在线免费观看| 亚洲第一av免费看| 日韩中字成人| av女优亚洲男人天堂| 色94色欧美一区二区| 又粗又硬又长又爽又黄的视频| 亚洲欧美日韩另类电影网站| 伦精品一区二区三区| 美女主播在线视频| 美女内射精品一级片tv| 中文字幕免费在线视频6| 亚洲无线观看免费| 一区二区三区精品91| 免费av中文字幕在线| 丰满饥渴人妻一区二区三| 日日摸夜夜添夜夜添av毛片| 亚洲av中文av极速乱| 久久青草综合色| 91成人精品电影| 边亲边吃奶的免费视频| 99九九线精品视频在线观看视频| 亚洲图色成人| 欧美日韩视频精品一区| 一本一本综合久久| 亚洲经典国产精华液单| 丰满乱子伦码专区| 乱人伦中国视频| 国产精品国产三级专区第一集| 久久久久人妻精品一区果冻| 欧美成人午夜免费资源| 男女边摸边吃奶| 丰满少妇做爰视频| 不卡视频在线观看欧美| 午夜久久久在线观看| 99热这里只有是精品50| 九九久久精品国产亚洲av麻豆| 亚洲精品中文字幕在线视频 | 只有这里有精品99| 成人无遮挡网站| 亚洲久久久国产精品| 精品午夜福利在线看| 欧美日韩亚洲高清精品| 22中文网久久字幕| 亚洲四区av| 看十八女毛片水多多多| 免费av中文字幕在线| 又爽又黄a免费视频| 日韩视频在线欧美| 欧美bdsm另类| 国产美女午夜福利| 欧美日韩综合久久久久久| 少妇裸体淫交视频免费看高清| videos熟女内射| 日日啪夜夜撸| 久久国内精品自在自线图片| 亚洲欧美日韩东京热| 久久影院123| 欧美日韩亚洲高清精品| 午夜福利影视在线免费观看| 国产成人午夜福利电影在线观看| 美女国产视频在线观看| 亚洲欧美一区二区三区国产| 久久久久久久久久成人| 一级a做视频免费观看| 99热全是精品| 国产白丝娇喘喷水9色精品| 一级片'在线观看视频| 国产乱来视频区| 边亲边吃奶的免费视频| 观看美女的网站| 乱人伦中国视频| 这个男人来自地球电影免费观看 | 伊人亚洲综合成人网| 久久韩国三级中文字幕| kizo精华| 精品久久久久久久久av| 乱人伦中国视频| 欧美日韩在线观看h| 国产黄色视频一区二区在线观看| 精品国产一区二区三区久久久樱花| 色视频在线一区二区三区| 精品视频人人做人人爽| 少妇人妻 视频| 麻豆精品久久久久久蜜桃| 亚洲精品视频女| 色婷婷久久久亚洲欧美| av.在线天堂| 嫩草影院入口| 成年人免费黄色播放视频 | 嘟嘟电影网在线观看| 亚洲精品,欧美精品| 麻豆乱淫一区二区| 国产精品偷伦视频观看了| 黄色视频在线播放观看不卡| av天堂中文字幕网| 国产精品久久久久久av不卡| 女性生殖器流出的白浆| 亚洲,一卡二卡三卡| 亚洲国产日韩一区二区| 亚洲av电影在线观看一区二区三区| 成人综合一区亚洲| 七月丁香在线播放| 中文字幕亚洲精品专区| 久久99精品国语久久久| 熟女人妻精品中文字幕| 免费观看在线日韩| 亚洲高清免费不卡视频| 人妻制服诱惑在线中文字幕| 国产日韩一区二区三区精品不卡 | av不卡在线播放| 免费av不卡在线播放| 最近中文字幕高清免费大全6| 人人妻人人爽人人添夜夜欢视频 | 精品一区二区三区视频在线| av福利片在线观看| 国产日韩一区二区三区精品不卡 | 一级毛片我不卡| 黄色一级大片看看| 一级毛片久久久久久久久女| 日韩欧美一区视频在线观看 | 如何舔出高潮| 亚洲精品久久久久久婷婷小说| 日韩亚洲欧美综合| 色网站视频免费| 91久久精品国产一区二区成人| 免费看不卡的av| 精品久久久久久久久av| 又黄又爽又刺激的免费视频.| 视频区图区小说| 丝瓜视频免费看黄片| 日韩精品有码人妻一区| 99热这里只有精品一区| 国产一区二区在线观看日韩| 中文字幕制服av| 精品久久久久久久久亚洲| 国产精品久久久久久久久免| 免费人妻精品一区二区三区视频| 亚洲精品日本国产第一区| 免费观看在线日韩| 亚洲欧美清纯卡通| av福利片在线观看| 99久久综合免费| 人妻制服诱惑在线中文字幕| 嫩草影院新地址| 久久久久久久久久成人| 免费大片18禁| 伦精品一区二区三区| 久久综合国产亚洲精品| 人人妻人人爽人人添夜夜欢视频 | 在现免费观看毛片| 日本免费在线观看一区| 王馨瑶露胸无遮挡在线观看| 国产白丝娇喘喷水9色精品| 久久99热这里只频精品6学生| 尾随美女入室| 视频中文字幕在线观看| 国产成人freesex在线| av在线播放精品| 少妇高潮的动态图| 日日摸夜夜添夜夜爱| 尾随美女入室| 校园人妻丝袜中文字幕| 欧美+日韩+精品| 色视频在线一区二区三区| 日韩,欧美,国产一区二区三区| 一区二区三区四区激情视频| 久久精品国产亚洲av涩爱| 国产高清有码在线观看视频| 黄色一级大片看看| 久久久久久久精品精品| 日本黄色日本黄色录像| 日日爽夜夜爽网站| 色94色欧美一区二区| 大片免费播放器 马上看| 波野结衣二区三区在线| 欧美日韩国产mv在线观看视频| 久久女婷五月综合色啪小说| 18禁在线无遮挡免费观看视频| 熟女电影av网| 五月天丁香电影| 日本免费在线观看一区| 最后的刺客免费高清国语| 在线看a的网站| 久久久久网色| 亚洲精品aⅴ在线观看| 久久久久精品性色| 一级,二级,三级黄色视频| 国产黄色视频一区二区在线观看| 边亲边吃奶的免费视频| 在线免费观看不下载黄p国产| 男女啪啪激烈高潮av片| 午夜激情福利司机影院| 日本av免费视频播放|