• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A graph-based two-stage classification network for mobile screen defect inspection*

    2023-03-04 15:03:42ChaofanZHOUMeiqinLIUSenlinZHANGPingWEIBadongCHEN

    Chaofan ZHOU ,Meiqin LIU ,Senlin ZHANG ,Ping WEI ,Badong CHEN

    1State Key Laboratory of Industrial Control Technology,Zhejiang University,Hangzhou 310027,China

    2College of Electrical Engineering,Zhejiang University,Hangzhou 310027,China

    3Institute of Artificial Intelligence and Robotics,Xi’an Jiaotong University,Xi’an 710049,China

    Abstract:Defect inspection,also known as defect detection,is significant in mobile screen quality control.There are some challenging issues brought by the characteristics of screen defects,including the following: (1) the problem of interclass similarity and intraclass variation,(2)the difficulty in distinguishing low contrast,tiny-sized,or incomplete defects,and (3) the modeling of category dependencies for multi-label images.To solve these problems,a graph reasoning module,stacked on a classification module,is proposed to expand the feature dimension and improve low-quality image features by exploiting category-wise dependency,image-wise relations,and interactions between them.To further improve the classification performance,the classifier of the classification module is redesigned as a cosine similarity function.With the help of contrastive learning,the classification module can better initialize the category-wise graph of the reasoning module.Experiments on the mobile screen defect dataset show that our two-stage network achieves the following best performances: 97.7% accuracy and 97.3% F-measure.This proves that the proposed approach is effective in industrial applications.

    Key words: Graph-based methods;Multi-label classification;Mobile screen defects;Neural networks

    1 Introduction

    With the mass production of mobile phone screens,accurate and efficient defect inspection is significant for mobile phone screen industrial processes.Manual inspection requires plenty of experienced inspectors,which is labor-consuming and inefficient.In addition,due to the special optical properties of mobile screens,long-term inspection does some harm to the eyes of inspectors.With the rapid development of image processing and deep learning technologies,many automated optical inspection solutions based on convolutional neural networks (CNNs) have been proposed in the textile industry (Wei et al.,2021),steel industry (Wang YC et al.,2021b),mobile phone screen manufacturing industry(Lei et al.,2018;Yuan et al.,2018;Li et al.,2020;Lu et al.,2020;Wang T et al.,2021;Zhang et al.,2021;Zhu et al.,2022),and so on.

    This study focuses on mobile screen defect inspection,and the images are obtained by an image acquisition device,as shown in Fig.1.The device is composed of a line scan camera,an annular light source,transmission parts,a controller,and a computer.Since the original mobile screen image is too large to process directly,we cut it into patches with the size of 1024×1024.The cropped image patches are displayed in Fig.2.

    Fig.1 Image acquisition device and an image obtained by the device

    Fig.2 Samples of mobile screen defects indicated by manual annotation: (a) original image;(b-f) images after adjusting the brightness for view,where defects that appear in the corresponding subfigures from left to right and top to bottom are bubble,scratch,pinhole,scratch (a,b),scratch,tin ash (c),tin ash (d),pinhole,pinhole,pinhole (e),and pinhole,bubble (f).Some defects have the characteristics of interclass similarity and intraclass variation (ISIV).Some defects are of low contrast or are tiny-sized or incomplete

    Although directly applying the CNN model to defect inspection can achieve better results than traditional approaches that use hand-crafted features,there are some challenging issues brought by taskspecific characteristics of mobile screen defect inspection.We show some samples of mobile screen defects in Fig.2.However,the original image is dark and inconvenient to see,as presented in Fig.2a.Therefore,the rest of the images in Fig.2 are brightened for easier viewing.First,some defects have the characteristics of interclass similarity and intraclass variation(ISIV).For example,the scratches in Figs.2b and 2c belong to the same category,but their appearances vary greatly.The scratch in Fig.2c and the bubble in Fig.2b are somewhat similar in shape.The local appearance of tin ash in Fig.2d is rather like the appearance of a pinhole,presented in Fig.2e.As a consequence of this characteristic,it is hard to classify defects correctly.Second,some defects on mobile screen images are of low contrast or are tiny-sized or incomplete.For example,the bubble in Fig.2f is not as complete as the bubble in Fig.2b.The pinhole in Fig.2f is too small and of too low contrast to be seen.These phenomena make it easy to miss defects during inspection.Third,we need to address the problem of multi-label classification.Compared with single-label classification,multi-label classification is a more practical and general problem (Wei et al.,2021),as a real-world image generally contains single or multiple defect classes,and so do the data we collect.As far as we know,there are a few research works (Lei et al.,2018;Li et al.,2020;Lu et al.,2020) on mobile screen defect classification up to now,but they study only single-label classification,which aims at classifying an image into one class.Different from them,we address the problem of multi-label classification on a mobile screen defect dataset.The multi-label task is more challenging,and a key to multi-label classification is to model the category dependencies.

    To address the above issues,a two-stage defect inspection network is proposed for multilabel classification of mobile phone screen defects.The proposed network consists of two components,i.e.,a feature-clustering-based classification module(FC2-module) and a graph reasoning module (GRmodule).To address the first problem,we create a category-to-category graph in the GR-module.The graph enhances the image features with the help of category features to increase the interclass distance and the intraclass correlation in the feature space.

    To obtain better node representations of the graph in the GR-module,we first use contrastive learning in the FC2-module to pull together features with the same label,while pushing apart features with different labels.Then,multiple category clusters are formed.However,contrastive learning cannot be directly applied to the multi-label problem;so,we propose the category pooling operator to generate label-level features.The classifier of the FC2-module is designed as a cosine similarity function to make the weights represent the overall features.We use the distilling weights as node representations.

    Besides,in the GR-module,we establish imagewise relations to improve features of images that have low contrast,tiny-sized,or incomplete defects.The use of image-wise relations is inspired by this phenomenon.When humans are not sure about the objects in the image being viewed,they refer to other images in which similar objects occur.Moreover,we model the category dependencies using the graph convolution network (GCN) and the co-occurrence patterns of defects.

    In summary,the major contributions of this paper are as follows:

    1.Based on a category graph and image relations,the GR-module exploits category-wise dependency,image-wise relations,and interactions between them to enhance original image features.The category correlation introduced in the image features is applied to increase the interclass distance and decrease the intraclass distance in the feature space.With image-wise relations,the features of low-quality images can be improved with reference to other high-quality images.Category dependencies embodied in the category graph help improve the multi-label classification performance.

    2.The FC2-module is designed to help the GRmodule better address the three problems mentioned above.For the better initialization of the graph of the GR-module,and consequently improving the multi-label classification performance,we propose a novel classification module (i.e.,FC2-module).This module uses contrastive learning and cosine similarity to obtain weights that contain enough category information for initialization.We design a simple category pooling operator to solve the dilemma of applying contrastive learning to multi-label defect inspection.

    3.Our experiments on a real-world industrial dataset demonstrate that the proposed two-stage classification network is effective in mobile screen defect inspection.

    2 Related works

    2.1 Defect inspection

    This subsection introduces mainly approaches for mobile screen defect inspection,graph-based defect classification,and multi-label defect classification.Nowadays,many CNN-based models have been widely used in classification(Lei et al.,2018;Li et al.,2020;Lu et al.,2020),object detection (Wang T et al.,2021;Zhang et al.,2021;Zhu et al.,2022),and segmentation (Yuan et al.,2018) for mobile screen defect inspection.These works demonstrate that CNNs can capture much richer features to overcome the limitations of traditional methods,i.e.,the methods applying hand-crafted features.However,different from our work,some approaches(Lei et al.,2018;Li et al.,2020;Lu et al.,2020) are aimed at solving the problem of single-label classification,while others (Yuan et al.,2018;Wang T et al.,2021;Zhang et al.,2021;Zhu et al.,2022) experiment only on single-label images.However,the images we face are multi-label.A key to multi-label classification is modeling category dependencies,which is not considered by these works.Besides,because only one or two defects are considered,or the task type is different,they do not research ISIV,the factor that is prone to cause wrong inspection.

    Some studies (Wang YC et al.,2021a,2021b;Kong et al.,2022;Xiao et al.,2022) introduced the GCN for surface defect classification.A semisupervised algorithm was proposed by Wang YC et al.(2021a).They used the GCN to propagate label information so that unlabeled images could obtain labels.Wang YC et al.(2021b)applied the GCN for defect classification and addressed the problem of class imbalance.Xiao et al.(2022) addressed the few-shot problem and introduced the graph to embed features of the support set into query features.Different from us,the graphs proposed by these authors(Wang YC et al.,2021a,2021b;Xiao et al.,2022)were image-to-image graphs and were applicable to only single-label classification,not multi-label classification.Kong et al.(2022) designed a knowledge graph but considered only the knowledge between bolt and nut pairs.They did not consider the relationship between the same categories.

    Some works (Park et al.,2020;Haurum and Moeslund,2021;Wei et al.,2021) do address the problem of multi-label defect classification.Nevertheless,the field and characteristic of defects they studied are different from those of ours.They studied defects in printed circuit board(PCB)screen printer,sewer,and textile,respectively.Directly applying these methods to mobile screen defects may degrade the classification performance.As shown in Section 4.3,the model (Haurum and Moeslund,2021)does not perform well on our dataset.Based on the characteristics of mobile screen defects,we propose a different method that exploits category-wise dependency,image-wise relations,and interactions between them.

    2.2 Graph-based multi-label classification

    Modeling the label dependencies is a key to help multi-label classification.There are quite a few works attempting to capture category dependencies by using probabilistic graph model,attention mechanism,GCN (Chen ZM et al.,2019;Wang Y et al.,2020;Zhao et al.,2021),and so on.Graph has been proved to be more effective in modeling category correlation (Chen ZM et al.,2019).Although these GCN-based methods have achieved appealing performance in natural image tasks,it is unwise to use them directly in industrial images.Some methods(Chen ZM et al.,2019;Wang Y et al.,2020)initialize graph nodes with linguistic word embeddings learned by language representation models.They ignore the characteristics of each image.Besides,training word embeddings for classifying industrial defects is hard.Different from them,we use a dynamically updated graph,whose node representations are initialized by the weights of the classifier.Additionally,other methods(Chen ZM et al.,2019;Wang Y et al.,2020;Zhao et al.,2021)neglect image-wise relations and image-to-category interactions,which we use to improve the classification performance.

    2.3 Contrastive learning

    Contrastive learning is proposed to pull together features with the same label while pushing apart features with different labels under self-supervised or fully-supervised settings.Under the fully-supervised setting,ground-truth label information is leveraged.Under the self-supervised setting,features generated from different images are deemed to have different labels.Recent works (Wu et al.,2018;Hjelm et al.,2019;Chen T et al.,2020;He et al.,2020;Khosla et al.,2020)have shown that contrastive learning can learn better representations and help improve performance in classification tasks.Contrastive learning under the self-supervised setting (Wu et al.,2018;Hjelm et al.,2019;Chen T et al.,2020;He et al.,2020)can work without labels,thereby relieving the burden of manual annotations.Nonetheless,for an image,these works (Wu et al.,2018;Hjelm et al.,2019;Chen T et al.,2020;He et al.,2020)took only augmented versions of the same image as positives,while methods under the fully-supervised setting(Khosla et al.,2020) also took images of the same class as positives.This principle causes the performance of some methods(Wu et al.,2018;Hjelm et al.,2019;Chen T et al.,2020;He et al.,2020)to be generally lower than the supervised contrastive method(Khosla et al.,2020).Consequently,in this study,the supervised contrastive method (Khosla et al.,2020) is introduced to learn better representations for our graph nodes.However,due to the multi-label attribute of our dataset,image features cannot be directly input to the supervised contrastive method(Khosla et al.,2020).To address this shortcoming,we introduce a new simple method to produce labellevel features of an image.Because the dimension of label-level features is low,the projection network in the original implementation is not required.

    3 Two-stage network and training policy

    In this section,we describe our two-stage network for mobile screen defect inspection and twophase training policy in detail.

    As depicted in Fig.3,the proposed two-stage network consists of two components,i.e.,the FC2-module and the GR-module.The former is able to classify defects and supply a good initialization for the graph of the latter,while the latter can be stacked on any existing classification network and improves the classification performance by reasoning semantics over image features with a category-wise graph.

    Fig.3 The proposed two-stage network.The input images are sent to the feature-clustering-based classification module (FC2-module) to obtain label-level features and distilling weights.Features and weights are fed into the graph reasoning module (GR-module) to enhance features.Finally,enhanced features are used to improve the classification performance

    Given a batch of input imagesI ∈RB×W×H×3,the FC2-module first uses a feature extractor and the category pooling operator to extractC × ddimensional featuresXas follows:

    whereBis the batch size,HandWdenote the height and width of the images respectively,andCmeans the number of categories.The pooling(·),referring to the category pooling operator,consists of a global average pooling layer and a linear layer.The features are fed into the first classifier to obtain classification scoresofBimages (wc1are learnable parameters of Cls1).Then,the featuresXare enhanced by reasoning semantics over image features with a category-to-category undirected graphG.Finally,the enhanced featuresX′are sent to the second classifier to obtain the final classification scoresS′=Cls2(X′|wc2)∈RB×C(wc2are learnable parameters of Cls2).To obtain better node features ofG,we implement the first classifier as a cosine similarity function between feature vectors and classification vectors,which is explained in detail in Section 3.1.

    3.1 FC2-module

    To make graphGbetter promote the model’s performance,we use the distilling weights of the first classifier as the representations of category nodes.To enhance the category representation capability of weights,the FC2-module is designed to contain three essential blocks: (1) a category pooling operator that changes image-level features to label-level features;(2)contrastive learning,improved based on multi-label characteristics,which gathers label-level features with the same label;(3) a classifier that makes the weights represent the overall features of each cluster.Next,the structure of this FC2-module is introduced in detail.

    Recently,the resurgence of the study of contrastive learning has brought about major advances in classification by learning better representations.To enhance the category representation capability of the weights of the first classifier,we add supervised contrastive loss to the FC2-module.Our dataset belongs to a multi-label dataset.Different from single-label classification based on contrastive learning (Khosla et al.,2020),multi-label classification is more complex.Image-level features extracted by the multi-label classification module cannot be used directly as label-level features.Hence,it is hard to define positive or negative samples for an anchor.We solve this problem by designing the category pooling operator.Taking the three-dimensional(3D)feature of an image as input,the category pooling operator uses a global average pooling layer to aggregate spatial information,and then applies a linear layer of sizeC×dto obtain featuresX ∈RCd,which can be seen as a set of label-level features

    The pipeline of updating the weights of the first classifier is depicted in Fig.4.Given a batch of input images,all label-level features can be seen asX={Xij ∈Rd|i ∈{1,2,...,B};j ∈{1,2,...,C}}and ground-truth labels can be expressed asY={Yij ∈{0,1}|i ∈{1,2,...,B};j ∈{1,2,...,C}}.

    Fig.4 The pipeline of updating the weights of the first classifier.This figure takes as an example the condition when the batch size is five and the number of categories is four.Four different types of borders in the features represent four different categories.The triangle in the cosine similarity space represents the weight of the classifier

    We defineAI={Xij ∈X|Yij=1}as a set that contains label-level features with active labels,andAD(i,j)=AI{Xij}as a set that contains all label-level features inAIexceptXij.We defineXijas the anchor.The positive set can be expressed asAP(i,j)={Xpj ∈AD(i,j)|Ypj=Yij=1},which contains label-level features with the same active labels asXijexceptXij.The negative set can be defined asAN(i,j)=AD(i,j)AP(i,j),which contains label-level features with different active labels.With the above notations,the supervised contrastive loss we use is written as shown in Eq.(2).In Eq.(2),τis a scalar temperature parameter and dist(·)refers to the distance function of contrastive loss.Previous research works (Gidaris and Komodakis,2018;Khosla et al.,2020) have proved that cosine similarity space can obtain better classification performance than just linear combination.Therefore,we apply the cosine similarity as the distance function

    of contrastive loss to form multiple category clusters.Furthermore,to make the weights of the classifier learn enough information about each defect category,we implement the first classifier as a cosine similarity function between feature vectors and weight vectors.Different from the conventional linear classifier,this classifier applies label-level features rather than image-level features as input and places features in the normalized space (i.e.,cosine similarity space)before outputting the classification scoresS.

    We apply binary cross-entropy (BCE) loss as the classification loss.In the cosine similarity space,the gradient of classification loss tends to encourage the weights of the first classifier to learn the overall features of each defect category.Finally,we obtain the updated weights that contain enough information about each defect category.

    3.2 GR-module

    3.2.1 Graph construction

    The GR-module is based on a category-tocategory undirected graphG=(N,E),whereN ∈RC×drefers to a set of category nodes andE ∈RC×Cis a set of edges encoding relations between corresponding categories.Inspired by current works (Gidaris and Komodakis,2018;Xu et al.,2019),in which the authors thought that the weights of the classifier actually contain high-level semantic information for each category,we use the distilling weights of the first classifier to be the representations of category nodes.The edges of graphGrefer to co-occurrence between corresponding categories.To constructE,we first count the occurrence of label pairs in the training set and obtain the statistical matrixM ∈RC×C.Then,row normalization is performed to obtain the adjacent matrixE:Eij=Finally,we setEas learnable parameters,which can be adaptively adjusted through iterations.

    3.2.2 Pipeline of the GR-module

    Based on graphG,this module exploits category-wise dependency,image-wise relations,and interactions between them.By doing so,the GRmodule can incorporate category semantics and image relations to guide enhancement of image features.Here,we explain the pipeline of the GR-module in detail.

    First,the GCN is instantiated and category representations are updated according to the prior knowledge graph edgesE,which is formulated as follows:

    Second,image-wise relations are modeled.This process can be described by the following two formulae:

    where Proj(·) is a projection function andi=1,2,...,B,j=1,2,...,B.In mobile screen images,some defects are of low contrast,tiny-sized,or incomplete.Detecting these defects is challenging because they have low resolution and limited information.The output resolution of CNN is usually much lower than the input resolution.This usually leads to lower resolution and lower discrimination of features corresponding to these defective regions.Therefore,the features of these images are not enough to represent some categories.This leads to the low scores of these categories.Furthermore,misclassification occurs.We hope that low-quality image features can be improved by exploiting image-wise relations.The application of image-wise relations is inspired by this phenomenon.When humans are not sure about the objects in the image being viewed,they refer to other images in which similar objects occur.Specifically,as in Eq.(4),we apply the projection function Proj(·) to map image features to a latent space for similarity computation.The projection function can also condense channels to reduce computation complexity.We instantiate the projection function Proj(·) as just a single linear layer.As in Eq.(5),we apply the dot-product operator and softmax operator to perform the similarity computation.Finally,the relations of images in the same batch are obtained.

    Third,the mapping between images and categories is established with the help of other images,formulated as follows:

    where Soft(·) is a softmax function.This function is used to normalize the scoresSover all categories.

    Next,new image features are generated under the guidance of category semantics and image relations.We concatenate the original image featuresXwith the new image features to produce the enhanced featuresX′.The process can be formulated as follows:

    wherewRNdenotes the set of learnable parameters,Fa(·) is an activation function,and Concat(·) refers to a concatenation function.The enhanced featuresX′have category relevance in the latter dimensions.This category correlation increases the interclass distance and decreases the intraclass distance in the feature space.

    Finally,the enhanced featuresX′are fed into the second classifier to obtain better classification results.The second classifier is implemented as a single linear layer without bias.We select BCE loss as the classification loss of the second classifier.

    3.3 Two-phase training policy

    Node information has an effect on the initial point of the model in the optimization space.It plays an important role in reinforcing the learning capabilities of the GR-module.However,during the initial training,the weights of the first classifier,which are used as representations ofN,are initialized randomly.The weights cannot contain high-level semantic information to express category representations.We can alleviate this problem by adjusting the proportion of the loss of the second classifier in the total loss.This can be implemented in many ways,including(1)progressively tuning up the proportion in predefined milestones and(2)designing a function for changing the proportion.However,experiments prove that they are time-consuming and hard to debug.Hence,we ultimately choose to pretrain the first classifier to obtain weights that contain enough category information.This is more efficient and easier to debug.In summary,the training procedure is split into two phases: the node pretraining phase and the normal training phase.We use the same training set as the input for both phases.

    In the first phase,we train only the FC2-module with classification lossLBCEand contrastive lossLSupCon.The parameters of the feature extractor and the weights of the first classifier are learned.The overall training loss of this phase can be expressed as follows:

    where the scalar parameterαcontrols the trade-offbetween classification loss and contrastive loss.

    After the weights of the first classifier are learned to contain enough high-level semantic information,we apply them to initialize the category nodes of graphG.Then,the training enters the second phase,during which we train the learnable parameters of the GR-module while we continue training the parameters of the FC2-module.The whole network is subjected to end-to-end training with the average sum of classification lossLBCEof the two classifiers.Algorithm 1 summarizes the whole training process of the proposed network.We empirically discover that the proposed two-phase training policy is effective.

    4 Experimental results and discussion

    4.1 Dataset and evaluation metrics

    We collect 80 mobile screen defect images with a size of 5120×10 240 from real industrial scenarios,where 56 images are for training,12 images are for validation,and 12 images are for testing.We cut patches with the size of 1024×1024 from the original images with a sliding window.The categories of patches are annotated by experienced inspectors.The patches whose defective area accounts for≤20%of the original defective area are not considered.The types of mobile screen defects include bubble,pinhole,scratch,and tin ash.Each image patch contains 0–4 types of defects.The number of patches of each category before applying data augmentation can be seen in Table 1.

    Table 1 Number of patches of each category in the training,validation,and testing sets

    Four metrics are adopted to evaluate the classification performance: accuracy,precision,recall,andF-measure.They are broadly used to evaluate the performance of the multi-label classification models.For each image,the predicted labels are deemed as positive when their confidence levels are >0.5.Of the four metrics,accuracy andF-measure are applied to evaluate the overall performance.

    4.2 Implementation details

    Our approach is implemented using the Py-Torch (Paszke et al.,2017) framework on a server with Ubuntu 16.04 operating system and four NVIDIA TITAN X graphics processing units(GPUs).ResNet-50(He et al.,2016)is used to implement the feature extractorF(·).During pretraining and normal training,the stochastic gradient descent(Bottou,2010) is used with a momentum of 0.9,a weight decay parameter of 0.0005,and a batch size of 32.We set the initial learning rate of both phases as 0.001.All iteration training is terminated after 40 epochs.The dimension of label-level featuresdis set to 512.Following Khosla et al.(2020),the temperature parameterτin Eq.(2) is 0.1.Moreover,αin Eq.(8) is set to 0.4.Data augmentation is commonly used to increase the diversity of the dataset,avoid overfitting of complex networks,and enhance the robustness of networks.We apply some data augmentations to the training set,including horizontal image flipping,vertical image flipping,rotation,and color jitter.The size of the training image is set as 1024×1024.

    4.3 Performance evaluation

    We compare our approach with state-of-the-art methods,including three prevalent CNN classification methods (ResNet-50 (He et al.,2016),VGG-16 (Simonyan and Zisserman,2015),and Inceptionv3 (Szegedy et al.,2016)),two graph-basedmulti-label classification methods (KSSNet (Wang Y et al.,2020) and TDRG (transformer-based dual relation graph) (Zhao et al.,2021)),and one multi-label defect inspection method (SewerMl(Haurum and Moeslund,2021)).Both KSSNet and TDRG construct category-wise relations using GCN to improve the performance of multi-label classification.The SewerMl mentioned by Haurum and Moeslund (2021)is a multi-label classification framework for defect inspection.

    Table 2 gives the results of the multi-label classification on the testing set.These metrics of classification performance are averaged over categories.It shows that our approach achieves the highest accuracy (97.7%),precision (97.5%),recall (97.0%),andF-measure (97.3%) on the mobile screen defect dataset.Compared with the second place of each metric,our model outperforms with a margin of 0.5%on accuracy,0.9% on precision,1.1% on recall,and 1.2% onF-measure.These results prove that the proposed method is feasible.TDRG and Inceptionv3 also achieve relatively good results.TDRG designs a category-wise graph and dynamically constructs the correlation matrix as graph edges in a learnable manner.This illustrates that exploring category relations based on the graph does help multi-label classification of mobile phone screen defects.Inception-v3 has a relatively large number ofnetwork branches,which allows the final extracted features to have different receptive fields.This may help it better identify defects of different sizes on the same image.

    Table 2 Comparisons with state-of-the-art methods on a multi-label screen defect classification task

    The per-class precision and recall of our approach are shown in Fig.5.Compared with other types of defects,the precision and recall for the defect “pinhole” are the lowest,which are 95.9% and 94.5%,respectively.The inspection performance of each category basically meets the needs of actual screen defect inspection.

    Fig.5 Per-class precision and recall of our approach on the mobile screen defect dataset

    4.4 Ablation studies

    4.4.1 Effects of training policy and components

    We perform ablation studies on the mobile screen defect dataset to verify the effectiveness of the GR-module,the node pretraining phase,and the cosine similarity space of the proposed approach.As illustrated in Table 3,the GR-module helps the proposed approach by reasoning semantics over image features with a category-to-category undirected graph.It increases accuracy by 0.8%andF-measure by 0.6%.The node pretraining phase can improve performance by 0.4%in terms of accuracy and 0.8%in terms ofF-measure.This indicates that it is important for our approach to improve the initial node features of the graphG.The node pretraining phase is beneficial in that point.

    To explore how much the cosine similarity space contributes to the improved results compared to the linear combination space,we conduct experiments.The only difference between the two kinds of spaces is that all the features of the former are normalized.Specifically,the cosine similarity space in Table 4 refers to the fact that we apply the cosine similarity as the distance function for contrastive learning,and the first classifier adopts the cosine similarity function.Linear combination space is implemented by using the dot-product distance function and the conventional linear classifier.As illustrated in Table 4,using cosine similarity space can obtain better classification performance than just applying linear combination space.

    Overall,the experimental results show that the training policy and the two components of the proposed approach all can improve the performance of the proposed approach.

    4.4.2 Parameter analysis

    In this subsection,we detail the experiments to show the influence of parameters and to find their most appropriate values.

    1.Parameterdof the category pooling operator

    The category pooling operator is applied to generate label-level features.The value ofddetermines the dimension of the label-level features and is criti-cal to the performance of the model.We experiment with different values ofd,and the results are shown in Table 5.Compared with other values,whendequals 512,the best results are achieved on almost all metrics.Fig.6a visualizes intuitively how performance changes with parameterd.We can see that the performance of the model is generally improved with the increase ofdat the beginning.When the value ofdexceeds 512,the performance of the model begins to deteriorate.According to Table 5 and Fig.6a,the most appropriate value ofdis 512,where the dimension of label-level features is enough to represent useful information without introducing too much noise.This might mean that the dimension of the label-level features is best set close to the dimension of the image-level features divided by the number of classes.

    Table 3 Ablation results of the GR-module and the node pretraining phase on the mobile screen defect dataset

    Table 4 Ablation results of cosine similarity space on the mobile screen defect dataset

    2.Parameterαin Eq.(8)

    Parameterαcontrols the contribution of contrastive learning to the total loss in the first phase.To explore how much contrastive learning actually contributes to the improved results,different values ofαare selected for the experiments.The results are presented in Table 6,and the results of accuracy andF-measure are plotted in Fig.6b.When contrastive learning is not applied or whenαis small,the label-level features are hard to cluster by categories,which prevents the weights of the first classifier fromlearning well.Whenαis too large,the contribution to classification loss is degraded too much,which prevents the weights from containing sufficient category information.From Table 6,we can see that the model can achieve better performance with the assistance of contrastive learning.The most appropriate value ofαis 0.4,which makes the proposed model acquire the best performance on almost all metrics.

    Table 6 Ablation results of different values of α on the mobile screen defect dataset

    Table 5 Ablation results of different values of d on the mobile screen defect dataset

    Table 6 Ablation results of different values of α onthe mobile screen defect dataset

    3.Batch sizeBin the second phase

    Image relation is an integral part of the loss of the GR-module,and the loss is computed only within batches.Consequently,the batch sizeBmay have effects on the multi-label classification results.In the spirit of fairness,we use the same model parameters pretrained in the first phase and alter the value ofBin the second phase.All experimental results are shown in Table 7,and the results of accuracy andFmeasure are depicted in Fig.6c.WhenBbecomes larger,the change trend of performance slows down;so,four decimal places are reserved in Table 7 and Fig.6c to present the change more clearly.We can observe that the performance of the proposed network is basically enhanced with the increase ofB.However,for the fairness of experiments,the batch sizeBof all models is set to 32.

    4.5 Method complexity

    Parameter counts and the number of FLoatingpoint OPerations (FLOPs) of different models with input image size of 1024×1024 are reported in Table 8.The former metric reflects the memory size of the model,and the “FLOPs” is used to measure the complexity of the model.Our proposed model has 2.469×107parameters and has about 8.610×1010FLOPs.Models that perform better,namely,ResNet-50,Inception-v3,TDRG,and our proposed model,have similar model complexity.Our proposed network needs about 4.5 h to train on the mobile screen defect dataset.In the real-world phone screen production line,the original image to be inspected is cropped into multiple patches with a sliding window.The image patches are detected in parallel with our proposed network.The results of the same original image are summarized to achieve classification and coarse location of defects.The proposed network is able to achieve the speed of 156 frames/s on image patches,which means that it can deal with about three original screen images per second.The inspection speed is acceptable in the real-world phone screen manufacturing industry.In the future,we will further optimize the code and model to meet the requirements of real-time inspection and high performance.

    Table 7 Ablation results of different values of B on the mobile screen defect dataset

    Table 8 Comparison of numbers of parameters and FLOPs of different methods with input image size of 1024×1024

    4.6 Network visualization

    For qualitative analysis,we show the visualization for feature maps of different methods by using the class activation maps (CAMs) (Zhou et al.,2016).CAM is a method for interpreting the prediction decision made by CNN-based models.As illustrated in Fig.7,the maps of our approach can highlight regions of low contrast or tiny-sized/incomplete defects and other illegible regions caused by the characteristic of ISIV.These regions are neglected by ResNet-50 (He et al.,2016).We give a few examples.Fig.7a shows the original image to be visualized.Figs.7b–7e are the visualized images of Fig.7a.Figs.7b and 7d both belong to the map of tiny-sized defect (i.e.,class “pinhole”),but they correspond to ResNet-50 and our approach,respectively.We can see that the activate area of Fig.7b is wider when compared with Fig.7d,thereby reducing the contribution of the defective information area to the prediction score.The data results also show that our model can better identify the two tiny defects.To be specific,the score of ResNet-50 is 0.02 and that of our approach is 0.83 for the defect “pinhole” of this image.Fig.7f shows the original image to be visualized.Figs.7g–7j are the visualized images of Fig.7f.Comparing Fig.7h with Fig.7j,the defect“scratch,”which is of low contrast and incomplete,is ignored by ResNet-50.Fig.7k shows the original image to be visualized.Figs.7l–7o are the visualized images of Fig.7k.As illustrated in Figs.7m and 7o(i.e.,the map of ResNet-50 and our approach),the defect“tin ash” is of too low contrast and different from others in the same class,resulting in positioning error of ResNet-50.In addition,it is clearly seen that the defective regions can be covered by the CAM masks of our approach.Therefore,our approach is quite helpful for capturing discriminative or meaningful regions by exploiting category-wise dependency,image-wise relations,and interactions between them.

    Fig.7 Visualization for feature maps with a class activation map (CAM): defective images and manual annotation (a,f,k);visualization results of ResNet-50 for the defect “pinhole” (b,g,l);visualization results of ResNet-50 for the defects “bubble” (c),“scratch” (h),and “tin ash” (m);visualization results of ours for the defect “pinhole” (d,i,n);visualization results of ours for the defects “bubble” (e),“scratch” (j),and “tin ash”(o)

    4.7 Generalization performance

    To prove the generalization performance of our proposed model,we perform experiments on the steel dataset (https://www.kaggle.com/competitions/severstal-steel-defect-detection/data) created by Severstal.This dataset,with 12 568 images in total,contains four types of steel surface defects.We choose this dataset because it is also a multi-label defect dataset.Each image may be defect-free,or may have one or more defects.The dataset is divided in the proportion 8:1:1 for training,validation,and testing.The division is available on this website(https://github.com/CFZ1/The-division-for-kagglesteel-dataset).All experimental results are shown in Table 9.Compared to other methods,our model achieves the best results on almost all metrics,which proves that it has good generalization capability.

    Table 9 Comparisons with the state-of-the-art methods on the steel surface defect dataset

    5 Conclusions

    Given the characteristics of mobile screen defect data from the real industrial pipelines and the diffi-culties in recognizing them,we proposed a two-stage defect inspection framework,consisting of the FC2-module and the GR-module.The former is able to classify defects,while the latter improves classification accuracy by reasoning semantics over image features with a category-to-category undirected graph.To solve the problem caused by ISIV of screen defects,we created a category-to-category undirected graph in the GR-module and used this graph to enhance image features.This category correlation in enhanced features increases the interclass distance and the intraclass correlation in the feature space.To obtain better node representations of the graph,we used contrastive learning and cosine similarity in the FC2-module.However,contrastive learning cannot be directly applied to the multi-label problem;so,we proposed the category pooling operator to solve it.Besides,the image-wise relations were exploited to improve low-quality image features.This helps the proposed network alleviate missed inspection caused by low contrast or tiny-size/incomplete defects.Moreover,we modeled the category dependencies by GCN and the co-occurrence patterns of defects.The experimental results on the mobile screen defect dataset showed that our model has better defect inspection performances: 97.7% accuracy and 97.3%F-measure.In addition,we visualized the feature maps of models with CAM,which showed that illegible defects can be detected by our approach.

    Contributors

    Chaofan ZHOU designed the research.Chaofan ZHOU and Meiqin LIU processed the data.Chaofan ZHOU drafted the paper.Senlin ZHANG helped organize the paper.Ping WEI and Badong CHEN revised and finalized the paper.

    Compliance with ethics guidelines

    Chaofan ZHOU,Meiqin LIU,Senlin ZHANG,Ping WEI,and Badong CHEN declare that they have no conflict of interest.

    Data availability

    Due to the nature of this research,participants of this study do not agree for mobile screen defect data to be shared publicly,so the supporting data are not available.The data used to validate the model generalization performance (i.e.,the steel surface defect data) can be found on the web links given in this paper.

    国产极品天堂在线| 丁香六月天网| 我的女老师完整版在线观看| 精品酒店卫生间| av视频免费观看在线观看| 乱人伦中国视频| 好男人视频免费观看在线| √禁漫天堂资源中文www| 色哟哟·www| 免费不卡的大黄色大毛片视频在线观看| 亚洲熟女精品中文字幕| 永久免费av网站大全| 亚洲精品久久久久久婷婷小说| 高清在线视频一区二区三区| 性色avwww在线观看| 插阴视频在线观看视频| 国产成人a∨麻豆精品| 一区二区三区乱码不卡18| 欧美日韩在线观看h| 一个人免费看片子| 三上悠亚av全集在线观看 | 精品亚洲成a人片在线观看| 少妇精品久久久久久久| 五月玫瑰六月丁香| 成年美女黄网站色视频大全免费 | 男人和女人高潮做爰伦理| 国产欧美日韩一区二区三区在线 | 精品人妻熟女毛片av久久网站| 不卡视频在线观看欧美| 内射极品少妇av片p| 亚洲性久久影院| 大香蕉久久网| 中文资源天堂在线| 精品人妻一区二区三区麻豆| 七月丁香在线播放| 观看av在线不卡| 国产一区二区三区av在线| 久久久午夜欧美精品| 91久久精品电影网| h日本视频在线播放| 日韩成人伦理影院| 高清在线视频一区二区三区| 欧美丝袜亚洲另类| 欧美3d第一页| 熟女电影av网| videos熟女内射| 亚洲国产成人一精品久久久| 亚洲欧美日韩另类电影网站| 午夜91福利影院| 日韩强制内射视频| 免费久久久久久久精品成人欧美视频 | 99热这里只有是精品在线观看| 亚洲综合精品二区| 精品少妇久久久久久888优播| 色哟哟·www| 国产精品人妻久久久影院| 哪个播放器可以免费观看大片| 国产免费福利视频在线观看| 国产毛片在线视频| 国产色婷婷99| 亚洲国产成人一精品久久久| 国产亚洲一区二区精品| 欧美 亚洲 国产 日韩一| 久久国产亚洲av麻豆专区| 99久久精品国产国产毛片| 能在线免费看毛片的网站| 久久 成人 亚洲| 亚洲va在线va天堂va国产| 好男人视频免费观看在线| 亚洲欧美清纯卡通| 嫩草影院入口| 水蜜桃什么品种好| 欧美日韩亚洲高清精品| 少妇精品久久久久久久| 久久久欧美国产精品| .国产精品久久| 男的添女的下面高潮视频| 如日韩欧美国产精品一区二区三区 | 男人添女人高潮全过程视频| 亚洲在久久综合| .国产精品久久| 大陆偷拍与自拍| 五月开心婷婷网| av网站免费在线观看视频| 国产精品成人在线| 美女国产视频在线观看| 日韩强制内射视频| 大香蕉97超碰在线| 国产精品久久久久久久久免| 欧美精品高潮呻吟av久久| 国产成人freesex在线| 国产成人精品久久久久久| 亚洲美女黄色视频免费看| 我的女老师完整版在线观看| 色视频在线一区二区三区| 国产精品嫩草影院av在线观看| 欧美3d第一页| h视频一区二区三区| 岛国毛片在线播放| 寂寞人妻少妇视频99o| 街头女战士在线观看网站| 十分钟在线观看高清视频www | 一边亲一边摸免费视频| 中文精品一卡2卡3卡4更新| 国产精品免费大片| 色视频在线一区二区三区| 亚洲成人手机| 亚洲婷婷狠狠爱综合网| 久久久久人妻精品一区果冻| 一边亲一边摸免费视频| 777米奇影视久久| 亚洲国产欧美在线一区| 丰满乱子伦码专区| 成人18禁高潮啪啪吃奶动态图 | 日本av免费视频播放| 深夜a级毛片| 国产精品秋霞免费鲁丝片| 国产成人91sexporn| 菩萨蛮人人尽说江南好唐韦庄| 国产高清国产精品国产三级| 18禁在线无遮挡免费观看视频| 精品人妻熟女av久视频| 久久精品国产a三级三级三级| 18+在线观看网站| 国产熟女欧美一区二区| 国产精品偷伦视频观看了| 九色成人免费人妻av| 99九九在线精品视频 | 在线观看免费日韩欧美大片 | 亚洲av电影在线观看一区二区三区| 最近手机中文字幕大全| 久久久久久久国产电影| 在线精品无人区一区二区三| 在线观看免费日韩欧美大片 | 99久久精品国产国产毛片| 国产伦精品一区二区三区四那| 自拍偷自拍亚洲精品老妇| 中文字幕av电影在线播放| 又爽又黄a免费视频| 高清黄色对白视频在线免费看 | 亚洲av二区三区四区| 18禁裸乳无遮挡动漫免费视频| 久久久亚洲精品成人影院| 18禁在线播放成人免费| 日本91视频免费播放| 久久免费观看电影| 一区二区三区四区激情视频| 欧美一级a爱片免费观看看| 国产极品天堂在线| 精品卡一卡二卡四卡免费| 日本欧美视频一区| 亚洲欧美精品专区久久| 我要看日韩黄色一级片| 国产精品99久久99久久久不卡 | 免费观看的影片在线观看| 老女人水多毛片| 久久人妻熟女aⅴ| av免费在线看不卡| 夜夜骑夜夜射夜夜干| 少妇精品久久久久久久| 午夜福利在线观看免费完整高清在| 国产成人免费观看mmmm| 2021少妇久久久久久久久久久| 亚洲国产毛片av蜜桃av| 欧美+日韩+精品| av.在线天堂| 少妇的逼好多水| 亚洲国产精品成人久久小说| 美女中出高潮动态图| 国产成人精品久久久久久| 国产一区亚洲一区在线观看| 麻豆成人午夜福利视频| 日韩欧美一区视频在线观看 | 欧美xxxx性猛交bbbb| 国产成人午夜福利电影在线观看| 三级国产精品欧美在线观看| 国产精品国产三级国产av玫瑰| 美女cb高潮喷水在线观看| 免费少妇av软件| 精品熟女少妇av免费看| av免费在线看不卡| 色视频在线一区二区三区| 人妻少妇偷人精品九色| 亚洲国产精品国产精品| 亚洲精品一二三| 男女边摸边吃奶| 久久ye,这里只有精品| 国产精品偷伦视频观看了| 99re6热这里在线精品视频| 男女啪啪激烈高潮av片| 一级毛片久久久久久久久女| 另类精品久久| 日本黄色日本黄色录像| 看非洲黑人一级黄片| 久久韩国三级中文字幕| 精品亚洲成国产av| 久久人妻熟女aⅴ| 高清av免费在线| 国产视频内射| 日韩,欧美,国产一区二区三区| 亚洲怡红院男人天堂| 老司机亚洲免费影院| kizo精华| 成人国产麻豆网| 成人无遮挡网站| 欧美xxxx性猛交bbbb| 亚洲精品乱码久久久v下载方式| 欧美激情国产日韩精品一区| 免费观看av网站的网址| 精品午夜福利在线看| 夜夜骑夜夜射夜夜干| 又爽又黄a免费视频| 国产精品一区二区在线观看99| 亚洲熟女精品中文字幕| 亚洲国产精品专区欧美| 一级二级三级毛片免费看| 国产免费一区二区三区四区乱码| 熟女人妻精品中文字幕| 少妇被粗大的猛进出69影院 | 少妇的逼好多水| 日韩 亚洲 欧美在线| 免费少妇av软件| 天天操日日干夜夜撸| 多毛熟女@视频| 亚洲欧美日韩另类电影网站| 久久99一区二区三区| 一区二区三区四区激情视频| 久久毛片免费看一区二区三区| 色婷婷av一区二区三区视频| 久热久热在线精品观看| 这个男人来自地球电影免费观看 | 男人狂女人下面高潮的视频| 青春草视频在线免费观看| 热re99久久精品国产66热6| 日韩欧美一区视频在线观看 | 有码 亚洲区| 亚洲国产精品成人久久小说| 欧美一级a爱片免费观看看| 成人毛片60女人毛片免费| 亚州av有码| 成人无遮挡网站| 国产美女午夜福利| 国产乱人偷精品视频| 久久久欧美国产精品| 久久99一区二区三区| 精品国产乱码久久久久久小说| 噜噜噜噜噜久久久久久91| 人妻少妇偷人精品九色| 亚洲第一av免费看| 人妻夜夜爽99麻豆av| 久久人人爽av亚洲精品天堂| 高清毛片免费看| 永久免费av网站大全| 久久人人爽av亚洲精品天堂| 免费黄网站久久成人精品| 亚洲经典国产精华液单| av免费在线看不卡| 简卡轻食公司| 国产淫语在线视频| 边亲边吃奶的免费视频| 国产精品免费大片| 一级黄片播放器| 国产成人午夜福利电影在线观看| 国产女主播在线喷水免费视频网站| 交换朋友夫妻互换小说| 99热国产这里只有精品6| 在线观看美女被高潮喷水网站| 最近2019中文字幕mv第一页| 国产色爽女视频免费观看| 伦理电影大哥的女人| 精华霜和精华液先用哪个| 欧美另类一区| 女性被躁到高潮视频| 夜夜爽夜夜爽视频| 中文字幕人妻丝袜制服| 精品久久久久久久久av| 日本黄色片子视频| 国产黄片美女视频| 色哟哟·www| 三上悠亚av全集在线观看 | 日韩大片免费观看网站| 99久久综合免费| 麻豆乱淫一区二区| 中文字幕久久专区| 女性生殖器流出的白浆| 黄片无遮挡物在线观看| 卡戴珊不雅视频在线播放| 观看免费一级毛片| 中文字幕制服av| 亚洲成人一二三区av| 亚洲国产日韩一区二区| 日韩制服骚丝袜av| 国产 精品1| av视频免费观看在线观看| 男人舔奶头视频| 99九九在线精品视频 | 国产欧美日韩综合在线一区二区 | 久久婷婷青草| 日韩av不卡免费在线播放| 久久ye,这里只有精品| 三上悠亚av全集在线观看 | 嫩草影院入口| 久久精品国产亚洲av天美| .国产精品久久| 成人黄色视频免费在线看| 国产精品一区二区在线不卡| 丰满人妻一区二区三区视频av| 亚洲怡红院男人天堂| 中文字幕精品免费在线观看视频 | 日本av免费视频播放| 国产亚洲最大av| 国内精品宾馆在线| 久久久精品免费免费高清| 国产精品久久久久久久电影| 久久久久久伊人网av| 免费观看无遮挡的男女| 建设人人有责人人尽责人人享有的| 国产一区有黄有色的免费视频| 成人亚洲精品一区在线观看| 午夜av观看不卡| 噜噜噜噜噜久久久久久91| 久久精品夜色国产| 80岁老熟妇乱子伦牲交| av网站免费在线观看视频| 久久久国产一区二区| 少妇 在线观看| 女人久久www免费人成看片| 国产亚洲91精品色在线| 欧美日韩亚洲高清精品| 欧美xxⅹ黑人| 99九九线精品视频在线观看视频| 久久久久精品性色| 久久久a久久爽久久v久久| 免费观看av网站的网址| 交换朋友夫妻互换小说| 哪个播放器可以免费观看大片| 国产高清国产精品国产三级| 日韩亚洲欧美综合| 欧美 亚洲 国产 日韩一| 高清毛片免费看| videos熟女内射| 欧美激情极品国产一区二区三区 | 熟女人妻精品中文字幕| 日本欧美视频一区| 在线观看av片永久免费下载| 国产成人91sexporn| 卡戴珊不雅视频在线播放| 国产一级毛片在线| 十八禁网站网址无遮挡 | 久久人人爽人人片av| 在线观看免费视频网站a站| 欧美日韩一区二区视频在线观看视频在线| 黑人高潮一二区| 免费观看无遮挡的男女| 精品亚洲成a人片在线观看| 一级毛片黄色毛片免费观看视频| 一级毛片 在线播放| 老司机影院成人| 大又大粗又爽又黄少妇毛片口| av.在线天堂| 精品久久久久久久久av| 国产片特级美女逼逼视频| 久久婷婷青草| 国产免费福利视频在线观看| 日韩大片免费观看网站| 国产淫片久久久久久久久| 国产一区有黄有色的免费视频| 免费观看av网站的网址| 爱豆传媒免费全集在线观看| 亚洲精品国产av成人精品| 91成人精品电影| 亚洲第一区二区三区不卡| 91久久精品国产一区二区三区| 狂野欧美激情性bbbbbb| 在线播放无遮挡| 少妇被粗大猛烈的视频| 插阴视频在线观看视频| 又黄又爽又刺激的免费视频.| 日韩人妻高清精品专区| 男女国产视频网站| 极品少妇高潮喷水抽搐| 国产黄色免费在线视频| 久久午夜福利片| 美女中出高潮动态图| 少妇人妻 视频| 少妇人妻精品综合一区二区| 国产精品.久久久| 亚洲精品乱码久久久v下载方式| 人人妻人人澡人人爽人人夜夜| 亚洲自偷自拍三级| 最近2019中文字幕mv第一页| 国产精品不卡视频一区二区| 在线观看免费日韩欧美大片 | 综合色丁香网| 午夜影院在线不卡| 国产一区二区三区av在线| 一区二区三区免费毛片| 久久av网站| 国产一区亚洲一区在线观看| 在线亚洲精品国产二区图片欧美 | 欧美精品一区二区免费开放| 五月天丁香电影| 免费av不卡在线播放| 9色porny在线观看| 久久国产精品男人的天堂亚洲 | 亚洲欧美成人精品一区二区| 中国三级夫妇交换| 精品99又大又爽又粗少妇毛片| 男人舔奶头视频| 99久国产av精品国产电影| 九草在线视频观看| 成人影院久久| 中文字幕久久专区| 菩萨蛮人人尽说江南好唐韦庄| 纯流量卡能插随身wifi吗| 亚洲怡红院男人天堂| 涩涩av久久男人的天堂| 免费久久久久久久精品成人欧美视频 | 搡女人真爽免费视频火全软件| 老熟女久久久| 97超视频在线观看视频| 午夜视频国产福利| 美女xxoo啪啪120秒动态图| 日韩在线高清观看一区二区三区| 亚洲色图综合在线观看| 国产熟女欧美一区二区| 国产美女午夜福利| 亚洲精品日韩av片在线观看| 丝瓜视频免费看黄片| 一本—道久久a久久精品蜜桃钙片| 国产高清三级在线| 午夜久久久在线观看| 少妇被粗大的猛进出69影院 | 深夜a级毛片| 国产爽快片一区二区三区| 亚洲av男天堂| 久热这里只有精品99| av天堂久久9| 日本黄色日本黄色录像| 亚洲va在线va天堂va国产| 国产成人精品无人区| 最近的中文字幕免费完整| 国产一区二区三区av在线| 自线自在国产av| 美女国产视频在线观看| 国产淫语在线视频| 成人18禁高潮啪啪吃奶动态图 | 大话2 男鬼变身卡| a级毛色黄片| 少妇丰满av| 在线观看人妻少妇| 亚洲精品国产成人久久av| 久久精品国产亚洲av涩爱| 嫩草影院入口| 天天操日日干夜夜撸| 亚洲一区二区三区欧美精品| 嫩草影院新地址| 黑丝袜美女国产一区| 精品一区二区三区视频在线| a级毛片在线看网站| 亚洲成人一二三区av| 国产伦精品一区二区三区视频9| 欧美精品高潮呻吟av久久| 精品一区在线观看国产| 人人妻人人看人人澡| 国产精品三级大全| 18+在线观看网站| 日日爽夜夜爽网站| 日本黄大片高清| 麻豆精品久久久久久蜜桃| 人人妻人人澡人人看| 在线观看一区二区三区激情| 成人影院久久| 久久女婷五月综合色啪小说| 丝袜脚勾引网站| 久久99精品国语久久久| 久久久久久久久久久丰满| 在线观看三级黄色| 麻豆成人av视频| 99九九在线精品视频 | 如何舔出高潮| 亚洲一级一片aⅴ在线观看| 少妇被粗大的猛进出69影院 | 一区二区av电影网| 日韩一区二区视频免费看| 在线观看美女被高潮喷水网站| 欧美最新免费一区二区三区| 国产黄色视频一区二区在线观看| 少妇丰满av| 国产免费又黄又爽又色| 在线观看美女被高潮喷水网站| 成人亚洲欧美一区二区av| 亚洲国产色片| av福利片在线| 午夜影院在线不卡| 午夜免费鲁丝| 永久网站在线| 国产亚洲一区二区精品| 一级毛片 在线播放| 国产成人午夜福利电影在线观看| 热re99久久国产66热| 男女免费视频国产| 免费看日本二区| 亚州av有码| 中国美白少妇内射xxxbb| videossex国产| 亚洲激情五月婷婷啪啪| 丝瓜视频免费看黄片| 国产精品三级大全| 伦理电影大哥的女人| 男人爽女人下面视频在线观看| 欧美日韩视频精品一区| 亚洲第一av免费看| 嫩草影院新地址| 亚洲欧美一区二区三区黑人 | 18禁裸乳无遮挡动漫免费视频| 97超碰精品成人国产| 午夜福利视频精品| 日韩av不卡免费在线播放| 熟妇人妻不卡中文字幕| 欧美国产精品一级二级三级 | 丝袜脚勾引网站| 少妇被粗大猛烈的视频| h视频一区二区三区| 高清不卡的av网站| 91久久精品电影网| 免费黄网站久久成人精品| 午夜久久久在线观看| 亚洲中文av在线| 国产精品一区www在线观看| 哪个播放器可以免费观看大片| 午夜影院在线不卡| 久久青草综合色| 人妻夜夜爽99麻豆av| 美女视频免费永久观看网站| 午夜激情久久久久久久| 国产成人精品福利久久| 国产美女午夜福利| 极品人妻少妇av视频| 亚洲精华国产精华液的使用体验| 一区二区三区免费毛片| 国产亚洲5aaaaa淫片| 日日啪夜夜爽| 成人影院久久| 中文字幕人妻丝袜制服| 91精品国产九色| 又粗又硬又长又爽又黄的视频| 久久精品国产亚洲av涩爱| 在线 av 中文字幕| 欧美日韩精品成人综合77777| 久久精品熟女亚洲av麻豆精品| 国产黄片视频在线免费观看| 永久网站在线| 纯流量卡能插随身wifi吗| √禁漫天堂资源中文www| 国产精品蜜桃在线观看| 男女边吃奶边做爰视频| 国产男女超爽视频在线观看| 国产精品国产三级国产专区5o| 成人国产av品久久久| 成人亚洲欧美一区二区av| 青春草国产在线视频| 日本wwww免费看| 欧美老熟妇乱子伦牲交| 欧美 亚洲 国产 日韩一| 国产熟女午夜一区二区三区 | 国产精品一区二区性色av| 午夜精品国产一区二区电影| 有码 亚洲区| 久久国产精品大桥未久av | 亚洲av男天堂| 亚洲久久久国产精品| 亚洲国产精品成人久久小说| 日本wwww免费看| 尾随美女入室| 国产亚洲精品久久久com| 国产一区二区在线观看日韩| 久久精品熟女亚洲av麻豆精品| 毛片一级片免费看久久久久| 亚洲精品视频女| 老女人水多毛片| av在线播放精品| 亚洲精品自拍成人| 国产色爽女视频免费观看| 国产高清有码在线观看视频| 午夜福利网站1000一区二区三区| 国产深夜福利视频在线观看| 18禁在线播放成人免费| 精品熟女少妇av免费看| 久久久久久久国产电影| 我要看日韩黄色一级片| 日本黄大片高清| 亚洲精品乱码久久久v下载方式| 亚洲av电影在线观看一区二区三区| 国产av码专区亚洲av| 建设人人有责人人尽责人人享有的| 自拍欧美九色日韩亚洲蝌蚪91 | 国产老妇伦熟女老妇高清| 久久精品久久久久久噜噜老黄| 亚洲精华国产精华液的使用体验| 国内精品宾馆在线| 一区二区av电影网| 久久久久久久久久成人| 水蜜桃什么品种好| 一本—道久久a久久精品蜜桃钙片| 亚洲国产日韩一区二区| 国产美女午夜福利| 18禁裸乳无遮挡动漫免费视频| 国产欧美亚洲国产| 人妻系列 视频| 少妇的逼好多水| 国产精品一区二区在线观看99| 成人二区视频| 男女国产视频网站| 男女免费视频国产|