• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Health Monitoring of Milling Tool Inserts Using CNN Architectures Trained by Vibration Spectrograms

    2023-02-17 03:11:56SonaliPatilSujitPardeshiandAbhishekPatange

    Sonali S.Patil,Sujit S.Pardeshi and Abhishek D.Patange

    Department of Mechanical Engineering,College of Engineering Pune,Pune,411005,India

    ABSTRACT In-process damage to a cutting tool degrades the surface finish of the job shaped by machining and causes a significant financial loss. This stimulates the need for Tool Condition Monitoring (TCM) to assist detection of failure before it extends to the worse phase. Machine Learning (ML) based TCM has been extensively explored in the last decade. However, most of the research is now directed toward Deep Learning (DL). The “Deep”formulation, hierarchical compositionality, distributed representation and end-to-end learning of Neural Nets need to be explored to create a generalized TCM framework to perform efficiently in a high-noise environment of cross-domain machining.With this motivation,the design of different CNN(Convolutional Neural Network)architectures such as AlexNet, ResNet-50, LeNet-5, and VGG-16 is presented in this paper. Real-time spindle vibrations corresponding to healthy and various faulty configurations of milling cutter were acquired.This data was transformed into the time-frequency domain and further processed by proposed architectures in graphical form,i.e.,spectrogram.The model is trained,tested,and validated considering different datasets and showcased promising results.

    KEYWORDS Milling tool inserts;health monitoring;vibration spectrograms;deep learning;convolutional neural network

    1 Introduction

    In subtractive machining processes,the desired surface finish of a job is achieved by the controlled removal of material [1]. Conventional methods such as turning, milling, and drilling are principally practiced processes in the industry[2].Most of the research attention has been directed towards the milling process owing to its versatile and wide-ranging applications [3]. Milling is an intermittent cutting operation performed using a controlled cutting tool configured with multiple inserts[4].Thus,in-process damage to a cutting tool degrades the surface finish of the job shaped by machining and causes a significant financial loss [5]. Therefore, the subject of tool condition monitoring is being persuaded through various approaches, such as diagnostic, prescriptive, predictive, as well as descriptive ones,and has emerged as a field worth investing in view of Industry 4.0[6].The predictive approach forecasts the possible failure of a cutting tool based on historic data and is helpful in preventive maintenance[7].Most of the study is inclined toward the prediction of tool wear through the design of deep networks[8].The diagnostic and prescriptive approach has been practiced for decades in an offline mode and is referred to as root cause analysis[9,10].The diagnostic approach clarifies the reason behind the problem and the prescriptive approach suggests that action needs to be taken to solve the problem[11,12].The descriptive approach characterizes the problem and the predictive approach helps in its prevention[13,14].Researchers have adapted a descriptive approach through ML models capable of characterizing the cutting tool condition[15,16].Li et al.[17]presented a comprehensive review based on structures of aero-engine rotors for analyzing their reliability.The surrogate models are recommended owing to their strong computational efficiency in terms of time and development procedure. Cherid et al. [18] presented optimization of the number of sensors and their placement in order to detect and localize the failure in bridge structure through the integration of Principle Component Analysis and Artificial Neural Net reduced FRF technique. Luo et al. [19] presented structural reliability analysis using the ML approach and hybrid novel Monte Carlo simulations for turbine bladed disk and laminated composite plate. This model showcases the greater flexibility in the prediction of failure over enhanced Monte Carlo simulation and it is efficient computationally with high-nonlinear problems.In an era of intelligent data analytics,the predictive approach has been reformed in an intelligent way by designing various ML and DL models capable of predicting the remaining useful life of a cutting tool[20].

    Most of the similar models being trained on exact and limited instances lack key features of generalization when deployed in a high noise environment of cross-domain machining and versatile process parameters. In order to provide descriptive analytics for similar situations, a deep learning framework is achieving substantial consideration and can be configured in desired ways. Once the approach is established, this can be further extended to diagnostic and prescriptive approaches in a similar manner. In the case of vibration signal processing, STFT is used for processing and analyzing signals of non-stationary nature by splitting them as segments of narrow time periods and taking each segment’s Fourier transform. As STFT has smaller time windows, subsequently,the spectrum of frequencies varies smoother over time series, therefore it appears more detailed.On the other hand, the model trained on directly sampled numeric vibration data may not yield real results. Thus, the STFT spectrogram breaks the time-domain response of vibration into the chain of folds and takes the Fast Fourier transform of all folds. This chain of FFT folds is then overlapped for visualizing how the magnitude and frequencies of the changes over a change in time.This transformation makes the 3D surface plot of FFTs and includes a scale of different colors and shades for representing and constructing a spectrogram image.A very first and important reason is that the deep learning algorithms work very efficiently on image-based datasets as they can be processed in multiple dimensions as compared to numeric datasets. Also, the information gained from images can be integrated into the time-series signals and has useful to make generalized robust models with fast convergence.Thus,classification using image-based datasets is advocated herein instead of using raw vibration signals in numeric form. Deep learning, a primary subclass of machine learning, has well established well for processing numeric as well as image-based data through the formulation of“Deep”structure,hierarchical compositionality,distributed representation,and end-to-end learning of Neural Nets.With this motivation,the design of different CNN(Convolutional Neural Network)architectures such as AlexNet, ResNet-50, LeNet-5, and VGG-16 is presented in this paper. Realtime spindle vibrations corresponding to healthy and various faulty configurations of milling cutter were acquired.This data was transformed into the time-frequency domain and further processed by proposed architectures in graphical form,i.e.,spectrogram.The model is trained,tested,and validated considering different datasets.The outline of the design,modeling,training and evaluation of CNN architecture algorithms is presented in Fig.1.

    Figure 1:Design,modeling,training and evaluation of classification algorithm

    2 Experimental Setup&Data Collection

    The experimental study is carried out in an industry named‘Axis Metal-cut Technologies’based in Pune, India. The experimental arrangement is represented in Fig.2. It is comprised of a Vertical Machine Center,a piezoelectric accelerometer(Make:PCB,Model:352C03 Piezotronics of+/-500 g range&10 mV/g sensitivity),and a data acquisition card(Vibration analyzer,Model:DEWE 43A with eight channels).The MS workpiece cuboid of 0.65 m×0.25 m×0.1 m was positioned using a fixture,and a milling cutter was held securely in the spindle. The milling cutter (diameter 63 mm) consisted of four carbide-coated inserts. With the intention to identify and study the fault classification, the standard machining parameters(cut depth,speed,and table feed)were chosen referring to the standard Komet catalog [21] and mathematical expressions. The speed at which the cutting is achieved was 180 m/min. The feed of the table was approximately 2000 mm/min. Finally, the depth of cut was preferred to be 0.25 mm as per the standard catalog ‘Kom-Guide 2016’ [21]. By considering these machining inputs and tool configuration,face milling operations were performed for 8 configurations,and change in vibration was collected. The 1sttool configuration labeled ‘HADF’(Healthy tool) was formed using defect-free,i.e.,normal inserts.The 2ndto 7thtool configurations labeled‘FFL,FNS,FNT,FCT,FTB,FBE’(faulty tool)were formed using defective inserts with flank wear,nose wear,notch wear,crater wear, tip breakage, built-up edge respectively introduced at one position of 4 positions of the tool inserts. The last configuration labeled ‘FAD’(faulty tool) was formed using all defective inserts(flank wear,nose wear,notch wear,crater wear)introduced at one all 4 positions.

    The vibration signals were acquired from the accelerometer considering the setting of sampling length,frequency,and the number of samples to be collected.A moderate sampling length of 8000 was considered.The frequency set for the sampling of vibration signal was 20 kHz according to Nyquist criteria.The Nyquist criteria state that the sampling frequency should be at least twice or more than the top frequency of the raw vibration signal. In this study, considering samples ranging from 50-100,the training data was created for each tool configuration such that it contains at least 2-3 lakh data points.Here 100 samples were considered for each configuration(class)of the milling cutter,and corresponding acceleration signals were collected using the DAQ framework and saved in the commadelimited data files.This forms training data set of a total of 800 samples.The disturbance caused by vibration produced by environmental conditions is filtered using the suitable low pass and high pass filters which were employed in the data acquisition process. In addition to this, the discrete wavelet transform can also be used to de-noise and decompose the original signal through different motherchild wavelet combinations. Once the data files are collected, signals can be read through graphical illustrations;thus accordingly their representation is essential and described in the next section.

    Figure 2:Experimental setup

    3 Experimental Setup&Data Collection

    The experimentation considering various configurations of milling tools has been performed to collect the data. It is indeed a challenging task to deal with raw data collected through dynamic experimentation. This necessitates the processing of signal that executes its synthesis, modification,and analysis. Signal processing mainly emphasizes detecting components of interest. Initially, the signal is read through graphical illustrations from different orientations.To commence with,the signal is represented in the time-domain and later adapted to spectrogram through the STFT, i.e., Shorttime Fourier Transform.The Fourier transform of aperiodic signals is actually the counterpart of the Fourier coefficients in periodic signals[22].The Discrete Fourier Transform(DFT)is a transform;it describes a relationship between a time sequence and a description of that sequence as a weighted sum of sampled data cosine and sine.The terms FFT and DFT are sometimes used interchangeably.But the DFT is the only one of the four that can be applied to digital data since it consists of a finite number of discrete samples [23]. The Discrete Time Fourier Transform (DTFT) is the dual of the Fourier series, an integral projection from continuous periodic frequency to sampled data time as opposed to an integral projection from continuous periodic time to the sampled data frequency. The DFT is a sample of the Discrete-Time Fourier Transform(DTFT).A transform is a change of variables we start with a variable usually‘t’for time and we change to a new variable usually omega for frequency.A series is not a change of variable it’s a different way to write the same function but the variable stays the same.The variable is still‘t’though despite the fact that the coefficients of the series are related to the function’s spectrum[24].A spectrogram is a graphical illustration of the variation of frequency with respect to time or simply depicts the strength of the vibration at a specific time instant.Owing to its graphical illustration,it appears in different colors and shades which is nothing but the measure of the strength of a signal. The bright shade indicates a signal with a higher strength which means the brightness of the shade has direct proportion with the strength of the signal. To summarize, a spectrogram is a fundamentally 2D graph, with a third dimension signified through colors [25-27].The time-frequency representation of vibration signal depicting different tool configurations is shown in Fig.3.The image data set consisted of 100 images of each of the 8 tool conditions which gave 800 images.

    Figure 3:Time-frequency representation of vibration signal depicting different tool configurations

    4 Mechanism of CNN Architectures

    CNNs are dedicated neural nets explicitly designed for capturing spatial(localized)facts within a data distribution. Through explicitly encoding them in architecture, CNNs are employed to train 2D images or 3D or even 1D data most of the time.The mechanism of CNN architectures(AlexNet,ResNet-50,LeNet-5,and VGG-16)is presented in this section.

    4.1 Lenet-5

    LeCun et al.[28]designed a pre-trained CNN model in 1998 which was introduced as Lenet-5 for document recognition considering Gradient-Based Learning.The architecture was specially designed for the recognition of printed and handwritten characters.The straightforward and simple architecture of this model was the key feature. The significance of ‘5’in Lenet-5 is that it has 5 layers with three convolution layers(CL)with averaging pools trailed by double fully connected layers(FCL)as shown in Fig.4. In the end, the Softmax classifier is connected that performs the classification [29]. The weighted sum (Aj) in classical neural nets aimed at the unitjis handed by the function ‘sigmoid squashing’for producing the condition of the unitjwhich is designated byXj

    The function‘sigmoid squashing’is formulated as follows which is originally a scaled-up hyperbolic tangent:

    Lastly,the output layer consists of Euclidean-RBF(Radial Basis Function),one for every label,having multiple input features.OutputYjis formulated as

    The criteria for sampling training samplers are as follows:

    Here,YDpis output and resembles the correct classification of input featuresZp.

    Figure 4:Basic architecture of Lenet[30]

    4.2 AlexNet

    At first, AlexNet is a significantly popular image classification algorithm that performs well on ImageNet. It’s one of the breakthrough algorithms in computer vision. It is not widely used nowadays, it’s been outperformed by other algorithms that are either less computationally intensive or have better performance on ImageNet[31].But still,a very interesting architecture to learn!Some of the significant points on AlexNet are discussed here.The AlexNet was split into two networks while training and trained in two GPUs,since it was too large for a single GPU to handle the network[32].

    It uses the ReLU activation function in place of the standard tanh function.Since AlexNet had 60 million parameters where ‘Dropout’and ‘Data augmentation’techniques were used to take care of the over fitting problem[33].The standard AlexNet architecture is shown in Fig.5.The neuron’s output is modeled by consideringfdesignates the function of the input featurexasf(x) =tanh(x)orf(x) = (1+e-x)-1.In the relationship of training time with respect to gradient descent,saturating nonlinearities is very slower than the non-saturating nonlinearityf(x) = max(0,x). The neurons with this nonlinearity as Rectified Linear Units(ReLUs)are preferred.Deep CNNs with ReLUs train much quicker than those of their counterparts withtanh.For case of the CIFAR-10 dataset,iterations desired to attain a training error of 0.25 using the 4-layered convolutional networks.This shows that large neural networks are essential to process such data.

    Figure 5:Basic architecture of AlexNet[34]

    4.3 ResNet-50

    ResNet is a CNN model that was developed to mitigate the issue of vanishing gradients[35].The key development is it introduces something called the identity shortcut connection which just skips a bunch of layers [36]. It can be seen as a special case of the Highway Network. ResNet Introduces a hyper-parameter called cardinality to adjust the model capacity and uses the split-transform-merge paradigm from the Inception model[37].XResNet applies a number of tricks that modify things like the training or model.Heuristics to increase the parallelism of training and decrease the computational cost through lower precision computing and modifying the learning rate or biases.

    Tweaking the models by modifying the network architecture is available. They explore several modifications they call ResNet A, ResNet B, ResNet C and ResNet D. These modify the stride length in particular convolutional layers. Training refinements assist in improving accuracy [35-37].ResNet-50 is a residual network where the number‘50’designates the no.layers involved in training.It is a subcategory of CNNs and is widely employed for the classification of images. The standard ResNet architecture is shown in Fig.6.Skipping connections is the novelty of ResNet[38].Without any hyper-parameters tuning, deep-oriented nets usually suffer through vanishing gradients as the algorithm takes back-propagation,the gradient reduces.A smaller gradient makes inflexible training.The identity function would be learned by the algorithm and allows the passing of the input features over the blocks without passing over the remaining weighted layers[39].This makes the stacking of extra layers and the network becomes deeper and deeper. Due to this vanishing gradient offsets by skipping the layers.Increasing the number of layers increases the nonlinearity of the neural network.That means, the problem becomes increasingly non-convex with an increasing number of layers.Highly non-convex planes are difficult to optimize and hence, the optimization algorithm will fail to find good minima. Residual Neural Networks solve this problem by introducing the idea of skip connections.ResNet introduces residual associates amongst layers,making the outcome of each layer a convolution of corresponding input along with its non-convolutional input[40].Furthermore,Batch Normalization is also used in the ResNet and incorporated into VGG architecture also. ResNet-50 is usually trained for big data, i.e., volume of which is measured in billion. The network consists of multiple layers and classifies images into different classes such as nose,notch,crater,flank wear,etc.Residual Network is popularly regarded as an efficient framework for training deep neural networks.Similar to a general Deep learning algorithm, the layers signify the nonlinear process of data for extracting features and transforming them to higher dimensions or lower dimensions.It also comprises a hidden layer of traditional ANN and an array of proposed formulae. Generally, in a deep CNN,stacking of layers can be found.The specialty of this net is that it learns features at several levels such as high/mid/low at the end of each layer.Also,in residual learning,as an alternative to trying to learn features,learning from some residual is desired.The concept of Residual is nothing but the deduction of learned features from non-learned features of a particular layer.ResNet makes it happen by shortcut connections,i.e.,by connecting theNthlayer to(N+X)thlayer directly.It is evident that training of such nets is simpler than conventional CNNs with superior accuracy[35-40].

    Figure 6:Basic architecture of ResNet[34]

    4.4 VGG-16

    The Visual Geometry Group(VGG)of the Department of Engineering Sciences at the University of Oxford designed the VGG-16 which is nothing but a novel trained Convolutional Neural Network architecture [40]. All the CNNs have more or less similar architecture, a stack of convolution and pooling layers at the start and ending with fully connected and soft-max layers.The‘VGG-16 Neural Network’was particularly trained for more than a million types of images with respect to the database of the‘ImageNet’and or the‘ImageNet’database respectively,where the network consists of various types of layers,and to be specific,there were 16 layers.The modification between the‘VGG-16’and the‘VGG-19’is that this type of network is 16 layers deep and that type of network was 19 layers deep respectively [41,42]. The ‘VGG Neural Net’was firstly presented by ‘Zisserman’and ‘Simonyan’in 2014 and was hence published in one of the research papers which was namely, ‘Large-Scale Image Recognition using Deep CNN’and was first published in this paper respectively.VGG-16 is just one of the configurations that won the 2014 ImageNet challenge in localization and classification tasks.16 indicate that the network has 16 weight layers as shown in Fig.7.VGG secures 1stand 2ndrank for the task of localizing and classifying for 2014’s challenge of ImageNet. All the CNNs have more or less similar architecture,the stack of convolution and pooling layers at the start and ending with fully connected and soft-max layers.The major role of VGG is to enhance accuracy by addition of depth of CNN despite smaller reception slices in earlier layers while performing the task of classifying/localizing images.

    Figure 7:Basic architecture of VGG-16[43]

    Nets previous to VGG were bigger reception slices,i.e.,11×11 and 7×7 in comparison with 3×3;however,are not as deeper as VGG.Some variants of VGG exist according to a number of weighted layers for example VGG 19. Suppose the input given to VGG 19 is 224×224 RGB image and the expected output is the probability value for each of 1000 classes of the ImageNet database.In this net,the Softmax layer creates an outcome by an application function named ‘softmax’as an activation function from the total input features received from the preceding layer. This function ‘softmax’compresses the outcome of every unit between 1 and 0,very similar to function‘sigmoid’.However,it also splits each outcome in a way that the total summation of the outcomes is 1. The outcome of this function is corresponding to discrete probability distributions and estimates the probability of true classes. Softmax is used in deep neural networks as an activation function at the output node.Maximum pooling put on pooling at maximum extent over its input features.Pooling is an operation in ConvNets,which combines the outcomes of a cluster of neurons at a single layer into one neuron in a succeeding layer.Maximum pooling analyzes every cluster of neurons at a previous layer and selects the maximum value from it.The application of VGG-16 is explained here.Suppose there is image data and the objective is to find the labels.The input to the network is an image of 224×224 ×3 sizes and there are 1000 images.There were convolutions(size 3×3)layers,max pooling(size 2×2)layers,and fully connected 16 layers at the end to make the size of the model size 528 MB.The 1strecommendation was to keep the size of all filters equal to 3×3;max pooling must be located after every 2 convolutions and filters number must be twice after every max-pooling.These recommendations were used to design the VGG-16 model for the current application[44].

    5 Design and Training of Architectures Based on Spectrogram Images

    CNN architectures (AlexNet, ResNet-50, LeNet-5, and VGG-16) under consideration were trained based on real-time spindle vibrations transformed as spectrogram images in order to develop a classification model for different tool configurations.

    ? The dataset consisted of 100 images of each of the 8 tool configurations,i.e.,in a total of 800 images.A total of 800 images were fed in different splits to each of the architecture.The splits range from 60% to 90% with the step of 10%. All four architectures were designed using the training set of split 60%,70%,80%,and 90%of 800 images.

    ? The hyper-parameters and architecture are given in Appendix A,B,C,and D respectively for all four architectures.The layers include convolution,max pooling,dense,and dropout.Other parameters are no.of filters,filter size,strides pool size,and perceptron.The activation function was ReLu and Softmax.

    ? The activation function keeps the outcome constrained to a specific kind and adds nonlinearity to data. The most popular functions are Gelu, Selu, Softsign, Softplus, Hyperbolic Tangent(tanh),Softmax,ELU,Sigmoid,Relu,etc.Sigmoid is usually employed for two-class categorization and can entail the problem of gradient vanishing.Further,it is not concentrated around zero and computation is very costlier. The tanh also causes the vanishing gradient problem. Swish, on the other hand, does not entail the problem of gradient vanishing and is established to be slightly superior to Relu however computation is complex. Relu is a better choice over Softplus and Softsign also.Relu is preferred for hidden layers.In the case of deep nets, the swish function works superior to Relu. Softmax Function is usually employed for multiple class categorizations and is thus located at the end nets. In the case of regression, a linear function is used at the end layer; for two class categorizations, sigmoid is the correct selection and for multiple class categorizations,softmax is the right choice[44].As this research investigates multiclass classification and the hidden layers are also used in the model.So ReLu and Softmax activation functions are used.

    ? Firstly, 60% of total images were employed to train all architectures. The trained model was then tested using 40%of the remaining data.Secondly,70%of the total images were used for training all four architectures and tested using 30%of the remaining data.Thirdly,80%of the total images were used for training all four architectures and tested using the remaining 20%of the remaining data.Lastly,90%of the total images were used for training all four architectures and tested using 10%of the remaining data.The number of epochs varied from 10 to 50.

    In the subsequent section,classification using spectrogram images has been discussed.

    While developing the AlexNet model, as shown in Table 1, the training was carried out using different splits as explained above.The highest accuracy of each split was found when the epoch was kept at 40. For training, the highest accuracy was obtained at a 70/30 split which was 94.69% and the lowest accuracy was achieved at a 90/10 split which was 87.86%. Since AlexNet split into two networks while training and trained in two GPUs since it was too large for a single GPU to handle the network.Thus for testing data,the highest accuracy was obtained at a 70/30 split which was 87.65%and the lowest accuracy was obtained at a 60/40 split which was 80.84%.For classifying blind data,the highest accuracy was obtained at 80/20 split which was 90.2%and the lowest accuracy was obtained at 90/10 split which was 81.6%.This shows that the average accuracy of classification was 86.98%.This shows no overfitting problem as AlexNet has nearly 60 million parameters where‘Dropout’&‘Data augmentation’takes care of overfitting.

    Table 1: Classification using AlexNet

    For the LeNet-5 algorithm,the highest accuracy of each split was found when the epoch was 50.It shows that for LeNet-5 from Table 2,with respect to training,the highest accuracy was obtained at 60/40,70/30 and 90/10 split which was 100%,and the lowest accuracy was achieved at 80/20 split which was 99.85%.Within just two convolution layers and two pooling layers followed by dense layers,the architecture overfits as the accuracy of training for all splits is almost 100%.Owing to the overfitting of the model in training,testing accuracy drops to 92%-93%.While classifying blind data,the highest accuracy was obtained at the 90/10 split which was 98.6%and the lowest accuracy was obtained at the 60/40 split which was 94.1%.

    Table 2: Classification using LeNet-5

    In the case of VGG-16,the highest accuracy of each split was found when the epoch was 20.It is apparent from Table 3,as there are 19 different layers;this model also overfits in training as accuracy is almost 100%.This overfitting in training drops testing accuracy to 87%-94%.For testing data,the highest accuracy was obtained at 80/20 split which was 94%and the lowest accuracy was obtained at the 60/40 split which was 87.44%.While classifying blind data,the highest accuracy was obtained at the 90/10 split which was 97.91%and the lowest accuracy was obtained at the 60/40 split which was 90%.

    Table 3: Classification using VGG-16

    The design of the ResNet-50 model shows that the highest accuracy of each split was found when the epoch was 20. The accuracy was substantially lower than other architectures despite rigorous hyperparameter tuning as shown in Table 4. The model mitigates the issue of vanishing gradients however the identity shortcut connection just skips a bunch of layers. Even though the heuristics increases the parallelism of training and decreases the computational cost but it affects precision computing and decreases the learning rate.This can be evident from the results.Concerning training,the accuracy ranges from 12.43%to 24.78%.Owing to insufficient training,testing accuracy also drops to range from 12.349%to 22.47%.While classifying blind data,the highest accuracy was obtained at the 80/20 split which was 19.3% and the lowest accuracy was obtained at the 60/40 split which was 14.89%. Overall comparison of all models shows that the highest average accuracy of classification(96.35%) was obtained through the LeNet-5 model, followed by VGG-16 (95.51%) and AlexNet(86.98%). The performance of ResNet-50 was very poor which could not classify faults more than 17.05%.In addition to this,the design of classifiers was examined with detailed results of classification.The validation results considering the dataset are presented in the next section with the help of confusion matrices to understand correct and incorrect classification of tool health.

    Table 4: Classification using ResNet-50

    6 Classification Results and Discussion

    The classification results, i.e., how designed CNN architecture classifies samples correctly are explained using the confusion matrix. The matrix is made up of 8 columns and 8 rows defining the actual and predicted class of tool conditions. In this section, classification results considering validation data are presented.After successful training and testing of models,for validation,out of 100 separate images that were created amongst each class,at random 36 photos were selected as the blind datasets.These 36 photos were then used to validate models trained using split 60%,70%,80%,and 90% and presented here. The confusion matrix for classification using AlexNet depicted in Fig.8d 90/10 is explained as follows. There were a total of 288 samples considering all labels. For all wear tool conditions,all 36 samples were correctly classified.For the built-up edge tool condition,only 13 samples were correctly classified out of a total of 36 samples;20 samples were misclassified as all wear,2 samples were misclassified as crater wear and 1 sample was misclassified as edge fracture.For the next case,crater wear condition,33 samples were correctly classified out of 36 samples;2 were misclassified as all wear and 1 was misclassified as healthy.Similarly,other classifications can be understood.

    Figure 8:Confusion matrix for classification using AlexNet

    The classification using LeNet-5 is depicted through confusion matrix in Figs.9a-9d. It can be seen that the classification of samples from‘a(chǎn)ll wear’,‘built up edge’,‘crater wear’,‘edge fracture’,and‘healthy tool’condition does not affect much over the change of % split of train and test. However,there is considerable change in the classification of notch, nose, and flank wear. The maximum misclassification occurs for ‘notch wear’ as it arises at the line of cutting depth and rubs over the work shoulder. This rubbing worsens and approaches tool surface abrasion and affects the cutting tool chemically which perhaps affects tool condition. The results for ‘notch wear’ in classification using AlexNet are even worst as only 9 samples were correctly classified in case of a 60/40 split.The confusion matrix for VGG-16 is shown in Figs.10a-10d. By carefully observing these matrices, the classification results for‘notch-wear’have been substantially improved in this case as compared to the previous case.The confusion matrix for the 80/20 split shows that 11 images were incorrectly classified and 277 images were classified accurately.Similarly,other classifications can be understood.Thus it is recommended that VGG-16 is useful to classify notch wear case as the abrasion on the tool surface is correctly recognized through multiple layers of VGG-16.Also,the classification of flank and nose wear conditions is superior to AlexNet.Overall the performance of VGG-16 and LeNet-5 is superior for all kinds of wears and is recommended for real-time implementation of these models.

    Figure 9:Confusion matrix for classification using LeNet-5

    Figure 10:Confusion matrix for classification using VGG-16

    7 Conclusions

    The use of deep learning-based techniques for mapping vibration response to the corresponding cutting tool condition has been successfully demonstrated from image-based data classification.The results obtained show that this methodology can effectively classify tool health conditions. Timefrequency vibrations response was observed for 8 different tool wears and has facilitated suitable tool condition classification.The CNN architecture-based classification methodology can be employed for the ideal use of cutting tools during manufacturing processes, to reduce idle times during manufacturing,and much more.The classification results achieved considering training,testing and blindfolds indicate that the methodology is proficient in tool health monitoring. Though the AlexNet model offers an accuracy of more than 85%, there is misclassification amongst faulty classes. This can be eliminated by considering the VGG-16 and LeNet-5 models where there was the least misclassification of flank,nose,and notch wear.The VGG-16 alone is sufficient to classify notch wear clearly.Overall the performance of VGG-16 and LeNet-5 is superior for all kinds of wears and thus recommended for real-time implementation of these models.Manual feature extraction was successfully eliminated with the help of this new deep network technology since manual feature extraction increases the computational time and complexity of the system. This system shall assist edge-old systems in up grading to be smarter with fewer resources.This scheme shall ensure the effective use of the cutting tool. The aforesaid strategies will be investigated in future research to decrease the amount of experimental effort necessary.It would be feasible to eliminate the requirement for data augmentation with a larger and more comprehensive dataset. This might also help with the development of a more precise classifier. Additionally, future research should be directed toward the deployment and real-time implementation of this methodology in the industry. With the recent advancements in MEMS-based sensors and instrumentation,the cost of data collection through experimentation has significantly reduced. For the case of measuring vibrations, MEMS-based accelerometers such as ADXL335/ADXL345 can be integrated with open source hardware such as Arduino/Raspberry Pi,and data can be stored/processed/displayed via Python/MS-Excel.

    Funding Statement:The authors received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    Appendix A.

    Hyperparameters for AlexNet

    Layers No.of filters Filter size Activation function Strides Pool size Perceptron Convolution layer 1 96 (11,11) ReLu 4 - -Batch normalization Max pooling - - - 2 (2,2) -Convolution layer 2 256 (5,5) ReLu 1 - -Batch normalization Max pooling - - - 2 (2,2) -Convolution layer 3 384 (3,3) ReLu 1 - -Convolution layer 4 384 (3,3) ReLu 1 - -Convolution layer 5 256 (3,3) ReLu 1 - -Batch normalization Max pooling - - - 2 (2,2) -Flatten Dense - - ReLu - - 4096 Dense - - ReLu - - 4096 Dropout 0.4 0.4 0.4 0.4 0.4 0.4 Dense - - Softmax - - 8

    Appendix B.

    Hyperparameters for LeNet-5

    Layers No.of filters Filter size Activation function Strides Pool size Perceptron Convolution layer 1 6 (3,3) ReLu 1 - -Batch normalization Max pooling - - - 2 (2,2) -Convolution layer 2 16 (5,5) ReLu 1 - -Batch normalization Max pooling - - - 2 (2,2) -(Continued)

    (continued)Layers No.of filters Filter size Activation function Strides Pool size Perceptron Flatten Dense - - ReLu - - 64 Dense - - ReLu - - 32 Dense - - Softmax - - 8

    Appendix C.

    Hyperparameters for VGG-16

    Layers No.of filters Filter size Activation function Strides Pool size Perceptron Identity block Convolution layer 1 F1 (1,1) ReLu 1 - -Batch normalization Convolution layer 2 F2 (1,1) ReLu 1 - -Batch normalization Convolution layer 3 F3 (1,1) ReLu 1 - -Batch normalization Convolutional block Convolution layer 1 F1 (1,1) ReLu s - -Batch normalization Convolution layer 2 F2 f ReLu 1 - -Batch normalization Convolution layer 3 F3 (1,1) ReLu 1 - -Batch normalization Convolution layer 4 F3 (1,1) ReLu s - -Batch normalization ResNet 50 architecture Convolution layer 1 64 (7,7) ReLu 2 - -Batch normalization(Continued)

    (continued)Layers No.of filters Filter size Activation function Strides Pool size Perceptron Max pooling - - - 2 (2,2) -Execution of convolutional block 1 for F1=64 F2=64 F3=256 f=3,3 s=1 Execution of identity block 1 for F1=64 F2=64 F3=256 f=3,3 Execution of identity block 2 for F1=64 F2=64 F3=256 f=3,3 Execution of convolutional block 2 for F1=128 F2=128 F3=512 f=3,3 s=2 Execution of identity block 3 for F1=128 F2=128 F3=512 f=3,3 Execution of identity block 4 for F1=128 F2=128 F3=512 f=3,3 Execution of identity block 5 for F1=128 F2=128 F3=512 f=3,3 Execution of convolutional block 3 for F1=256 F2=256 F3=1024 f=3,3 s=2 Execution of identity block 6 for F1=256 F2=256 F3=1024 f=3,3 Execution of identity block 7 for F1=256 F2=256 F3=1024 f=3,3 Execution of identity block 8 for F1=256 F2=256 F3=1024 f=3,3 Execution of identity block 9 for F1=256 F2=256 F3=1024 f=3,3 Execution of identity block 10 for F1=256 F2=256 F3=1024 f=3,3 Execution of convolutional block 4 for F1=512 F2=512 F3=2048 f=3,3 s=2 Execution of identity block 11 for F1=512 F2=512 F3=2048 f=3,3 Execution of identity block 12 for F1=512 F2=512 F3=2048 f=3,3(Continued)

    (continued)Layers No.of filters Filter size Activation function Strides Pool size Perceptron Max pooling - - - 2 (2,2) -Flatten Dense - - Softmax - - 8 Appendix D.Hyperparameters for ResNet-50 Layers No.of filters Filter size Activation function Strides Pool size Perceptron Identity block Convolution layer 1 F1 (1,1) ReLu 1 - -Batch normalization Convolution layer 2 F2 (1,1) ReLu 1 - -Batch normalization Convolution layer 3 F3 (1,1) ReLu 1 - -Batch normalization Convolutional block Convolution layer 1 F1 (1,1) ReLu s - -Batch normalization Convolution layer 2 F2 f ReLu 1 - -Batch normalization Convolution layer 3 F3 (1,1) ReLu 1 - -Batch normalization Convolution layer 4 F3 (1,1) ReLu s - -Batch normalization ResNet 50 architecture Convolution layer 1 64 (7,7) ReLu 2 - -Batch normalization Max pooling - - - 2 (2,2) -(Continued)

    (continued)Layers No.of filters Filter size Activation function Strides Pool size Perceptron Execution of convolutional block 1 for F1=64 F2=64 F3=256 f=3,3 s=1 Execution of identity block 1&2 for F1=64 F2=64 F3=256 f=3,3 Execution of convolutional block 2 for F1=128 F2=128 F3=512 f=3,3 s=2 Execution of identity block 3,4&5 for F1=128 F2=128 F3=512 f=3,3 Execution of identity block 4 for F1=128 F2=128 F3=512 f=3,3 Execution of convolutional block 3 for F1=256 F2=256 F3=1024 f=3,3 s=2 Execution of identity block 6-10 for F1=256 F2=256 F3=1024 f=3,3 Execution of convolutional block 4 for F1=512 F2=512 F3=2048 f=3,3 s=2 Execution of identity block 11&12 for F1=512 F2=512 F3=2048 f=3,3 Max pooling - - - 2 (2,2) -Flatten Dense - - Softmax - - 8

    久久午夜亚洲精品久久| 日日干狠狠操夜夜爽| 美女高潮喷水抽搐中文字幕| 亚洲熟妇中文字幕五十中出| 国产国拍精品亚洲av在线观看| 亚洲av成人av| 亚洲av一区综合| 欧美乱妇无乱码| 亚洲男人的天堂狠狠| 欧美激情国产日韩精品一区| 亚洲精品在线观看二区| 精品午夜福利在线看| 欧美在线一区亚洲| 亚洲avbb在线观看| 国产精品一区二区三区四区免费观看 | 桃色一区二区三区在线观看| 国产av麻豆久久久久久久| 欧美三级亚洲精品| 亚洲18禁久久av| 久久久久国内视频| 欧洲精品卡2卡3卡4卡5卡区| 看十八女毛片水多多多| 国内精品美女久久久久久| 夜夜看夜夜爽夜夜摸| 在线观看av片永久免费下载| 热99在线观看视频| 欧美一区二区国产精品久久精品| bbb黄色大片| 欧洲精品卡2卡3卡4卡5卡区| 日本成人三级电影网站| 久久中文看片网| 成人无遮挡网站| 亚洲美女搞黄在线观看 | 天堂av国产一区二区熟女人妻| 成年女人看的毛片在线观看| 国产精品一区二区三区四区久久| 亚洲成av人片免费观看| 99热精品在线国产| 91午夜精品亚洲一区二区三区 | 亚洲欧美清纯卡通| 国产在线精品亚洲第一网站| 成年女人看的毛片在线观看| 国产av不卡久久| 日韩亚洲欧美综合| av专区在线播放| 人妻夜夜爽99麻豆av| 丰满人妻一区二区三区视频av| 一边摸一边抽搐一进一小说| 淫妇啪啪啪对白视频| 国产亚洲精品久久久久久毛片| 草草在线视频免费看| 亚洲国产精品成人综合色| 亚洲欧美精品综合久久99| 国产精品99久久久久久久久| 国内揄拍国产精品人妻在线| 欧美xxxx黑人xx丫x性爽| 国产av一区在线观看免费| 一级黄色大片毛片| www.色视频.com| 久久天躁狠狠躁夜夜2o2o| 精品99又大又爽又粗少妇毛片 | 九九久久精品国产亚洲av麻豆| 性插视频无遮挡在线免费观看| 高清毛片免费观看视频网站| 国内久久婷婷六月综合欲色啪| 亚洲经典国产精华液单 | 国内精品久久久久精免费| 国产一级毛片七仙女欲春2| 精品一区二区免费观看| 乱人视频在线观看| 乱人视频在线观看| 三级男女做爰猛烈吃奶摸视频| 1000部很黄的大片| 精品久久久久久久久亚洲 | 成年版毛片免费区| 日本黄色视频三级网站网址| 日本黄大片高清| 亚洲精品在线观看二区| 久久精品国产亚洲av香蕉五月| 黄片小视频在线播放| 国产精品精品国产色婷婷| 欧美成人一区二区免费高清观看| 精品无人区乱码1区二区| 欧美国产日韩亚洲一区| 赤兔流量卡办理| 亚洲最大成人中文| 色哟哟哟哟哟哟| 一a级毛片在线观看| 久久亚洲真实| 精品久久久久久久久久久久久| 在线a可以看的网站| 国产私拍福利视频在线观看| 1024手机看黄色片| 亚洲成av人片在线播放无| 国产真实伦视频高清在线观看 | 老司机福利观看| 黄色日韩在线| 一二三四社区在线视频社区8| 久久久国产成人免费| 可以在线观看毛片的网站| 午夜福利高清视频| 观看美女的网站| 婷婷丁香在线五月| 久久香蕉精品热| 日韩欧美三级三区| 久99久视频精品免费| 脱女人内裤的视频| 国产毛片a区久久久久| 精品久久久久久久末码| 夜夜爽天天搞| 久久久久久久精品吃奶| av视频在线观看入口| 久久亚洲真实| 久久久久性生活片| 性欧美人与动物交配| 久久久久久久亚洲中文字幕 | 宅男免费午夜| 嫩草影院精品99| 日韩高清综合在线| 美女 人体艺术 gogo| 乱码一卡2卡4卡精品| 熟妇人妻久久中文字幕3abv| 亚洲成人免费电影在线观看| 精品人妻视频免费看| 欧美精品国产亚洲| 欧美在线一区亚洲| 国内精品久久久久久久电影| 精品99又大又爽又粗少妇毛片 | 亚洲av不卡在线观看| 天天一区二区日本电影三级| eeuss影院久久| 午夜亚洲福利在线播放| 一本久久中文字幕| 日本在线视频免费播放| av专区在线播放| 小说图片视频综合网站| 久久久久久九九精品二区国产| 最新中文字幕久久久久| 变态另类丝袜制服| 十八禁人妻一区二区| 毛片女人毛片| 一卡2卡三卡四卡精品乱码亚洲| 美女被艹到高潮喷水动态| 亚洲色图av天堂| 自拍偷自拍亚洲精品老妇| 久久人人爽人人爽人人片va | 久久伊人香网站| 欧美乱色亚洲激情| 大型黄色视频在线免费观看| 人人妻人人澡欧美一区二区| 久久久成人免费电影| 亚洲av熟女| 老女人水多毛片| 欧美一区二区精品小视频在线| 久久久精品欧美日韩精品| 亚洲av不卡在线观看| 亚洲国产高清在线一区二区三| 国产精品日韩av在线免费观看| 久久草成人影院| 免费av不卡在线播放| 国产高清视频在线观看网站| 首页视频小说图片口味搜索| 亚洲第一区二区三区不卡| 亚洲无线在线观看| 亚洲第一电影网av| 婷婷精品国产亚洲av在线| а√天堂www在线а√下载| 级片在线观看| 亚洲真实伦在线观看| 亚洲国产精品成人综合色| 精品一区二区三区av网在线观看| 日韩欧美精品v在线| 国模一区二区三区四区视频| 欧美日本视频| 日韩欧美三级三区| 亚洲欧美精品综合久久99| 亚洲 欧美 日韩 在线 免费| 一级毛片久久久久久久久女| 国产色婷婷99| 中文字幕人成人乱码亚洲影| 国产精品美女特级片免费视频播放器| 亚洲成av人片在线播放无| 少妇熟女aⅴ在线视频| 午夜福利在线观看免费完整高清在 | 国产亚洲av嫩草精品影院| 日韩高清综合在线| www.999成人在线观看| 九色成人免费人妻av| 一个人观看的视频www高清免费观看| 国产精华一区二区三区| 欧美极品一区二区三区四区| 亚洲av成人av| 乱人视频在线观看| 麻豆国产av国片精品| 日本三级黄在线观看| 国产aⅴ精品一区二区三区波| 伊人久久精品亚洲午夜| 国产极品精品免费视频能看的| 久久婷婷人人爽人人干人人爱| 国产精品精品国产色婷婷| 可以在线观看的亚洲视频| 欧美另类亚洲清纯唯美| 色5月婷婷丁香| 人人妻,人人澡人人爽秒播| 亚洲av美国av| 亚洲精华国产精华精| 午夜福利成人在线免费观看| 在线播放无遮挡| ponron亚洲| 女生性感内裤真人,穿戴方法视频| 国产麻豆成人av免费视频| a级一级毛片免费在线观看| 久久精品国产99精品国产亚洲性色| 少妇被粗大猛烈的视频| 女生性感内裤真人,穿戴方法视频| 观看美女的网站| 久久久色成人| 国产欧美日韩精品一区二区| 神马国产精品三级电影在线观看| 亚洲性夜色夜夜综合| 色视频www国产| 欧美激情国产日韩精品一区| 久久99热这里只有精品18| 男插女下体视频免费在线播放| 国产av麻豆久久久久久久| 麻豆成人av在线观看| 国产精品av视频在线免费观看| 一个人观看的视频www高清免费观看| 如何舔出高潮| 欧美不卡视频在线免费观看| 男女做爰动态图高潮gif福利片| 国产精品久久视频播放| 国内精品久久久久久久电影| 熟妇人妻久久中文字幕3abv| 久久婷婷人人爽人人干人人爱| 午夜老司机福利剧场| 午夜日韩欧美国产| 亚洲精品在线观看二区| 深爱激情五月婷婷| 久久精品国产99精品国产亚洲性色| 欧美日本视频| 欧美日本亚洲视频在线播放| 亚洲av免费在线观看| 18禁裸乳无遮挡免费网站照片| 人妻丰满熟妇av一区二区三区| av天堂在线播放| 露出奶头的视频| 少妇人妻精品综合一区二区 | 久久6这里有精品| 欧美成人一区二区免费高清观看| 午夜精品久久久久久毛片777| 国产69精品久久久久777片| 国产精品久久久久久久久免 | 亚洲va日本ⅴa欧美va伊人久久| 亚洲av二区三区四区| 亚洲精品在线美女| 国内少妇人妻偷人精品xxx网站| 亚洲成人精品中文字幕电影| 丰满的人妻完整版| 国产三级黄色录像| 99国产精品一区二区蜜桃av| 久久九九热精品免费| 日本黄色视频三级网站网址| 午夜福利18| 18禁在线播放成人免费| 国产亚洲欧美98| 亚洲专区国产一区二区| 日韩欧美三级三区| 99在线视频只有这里精品首页| 中文字幕av在线有码专区| 午夜福利欧美成人| 日本 欧美在线| 淫妇啪啪啪对白视频| 免费在线观看影片大全网站| 一二三四社区在线视频社区8| www.www免费av| 日韩大尺度精品在线看网址| 亚洲成人久久性| 亚洲国产欧美人成| 能在线免费观看的黄片| 免费av观看视频| 亚洲第一区二区三区不卡| 精品乱码久久久久久99久播| 十八禁人妻一区二区| 又爽又黄a免费视频| 欧美精品啪啪一区二区三区| 国产真实伦视频高清在线观看 | 内射极品少妇av片p| 日本黄大片高清| 久久精品影院6| 日韩亚洲欧美综合| 精品久久久久久久久久免费视频| 老司机深夜福利视频在线观看| 国内精品美女久久久久久| 亚洲av日韩精品久久久久久密| 国产精品亚洲av一区麻豆| 国产精品99久久久久久久久| 深夜a级毛片| 亚洲欧美日韩卡通动漫| 久久伊人香网站| 人妻夜夜爽99麻豆av| 最近视频中文字幕2019在线8| 婷婷丁香在线五月| 三级男女做爰猛烈吃奶摸视频| 有码 亚洲区| 国产精品久久久久久久久免 | 精品乱码久久久久久99久播| 老司机午夜十八禁免费视频| 伦理电影大哥的女人| 亚洲熟妇中文字幕五十中出| 少妇人妻精品综合一区二区 | 欧美区成人在线视频| 亚洲自偷自拍三级| 51国产日韩欧美| 好看av亚洲va欧美ⅴa在| 午夜福利欧美成人| 国产精品嫩草影院av在线观看 | 亚洲欧美日韩高清在线视频| 又粗又爽又猛毛片免费看| 丝袜美腿在线中文| 99热精品在线国产| 亚洲欧美日韩无卡精品| 亚洲精品久久国产高清桃花| 男女那种视频在线观看| 久久欧美精品欧美久久欧美| 中国美女看黄片| 午夜免费男女啪啪视频观看 | 又黄又爽又刺激的免费视频.| 人人妻,人人澡人人爽秒播| 成人午夜高清在线视频| 一本精品99久久精品77| 国产精品久久久久久亚洲av鲁大| 听说在线观看完整版免费高清| 久久久精品欧美日韩精品| 国产精品永久免费网站| 亚洲精品亚洲一区二区| 久久久久国内视频| 9191精品国产免费久久| 久久这里只有精品中国| 亚洲五月天丁香| 偷拍熟女少妇极品色| 99热这里只有是精品在线观看 | 精品人妻1区二区| 性插视频无遮挡在线免费观看| 国产高清三级在线| 老熟妇仑乱视频hdxx| 亚洲精品粉嫩美女一区| 欧美xxxx性猛交bbbb| 国产免费男女视频| 精华霜和精华液先用哪个| a在线观看视频网站| 日韩欧美国产在线观看| 免费在线观看影片大全网站| 免费看美女性在线毛片视频| av女优亚洲男人天堂| 色av中文字幕| 欧美三级亚洲精品| 十八禁网站免费在线| 91久久精品国产一区二区成人| 国产69精品久久久久777片| 久久久久久久久久成人| av在线观看视频网站免费| 一个人免费在线观看的高清视频| 欧美黄色淫秽网站| 日本五十路高清| 少妇人妻精品综合一区二区 | 久久久国产成人免费| 国产精品久久久久久久电影| 成年人黄色毛片网站| 亚洲精品在线美女| 别揉我奶头~嗯~啊~动态视频| 1000部很黄的大片| 无人区码免费观看不卡| 久久久久亚洲av毛片大全| 中文字幕精品亚洲无线码一区| 大型黄色视频在线免费观看| 日本撒尿小便嘘嘘汇集6| 日本五十路高清| 久久99热6这里只有精品| 人人妻,人人澡人人爽秒播| 亚洲人成伊人成综合网2020| 国产三级中文精品| 亚洲成a人片在线一区二区| 婷婷精品国产亚洲av在线| 久久久久性生活片| 国产三级中文精品| 日本黄色片子视频| 国产一区二区在线观看日韩| 如何舔出高潮| 久久99热这里只有精品18| 又爽又黄a免费视频| 国产私拍福利视频在线观看| 啦啦啦观看免费观看视频高清| 长腿黑丝高跟| 18+在线观看网站| 俄罗斯特黄特色一大片| 又黄又爽又免费观看的视频| 97热精品久久久久久| 九九在线视频观看精品| 99在线视频只有这里精品首页| 日韩欧美 国产精品| 国产人妻一区二区三区在| 国内少妇人妻偷人精品xxx网站| 一进一出抽搐动态| 日本五十路高清| 性欧美人与动物交配| 精华霜和精华液先用哪个| 校园春色视频在线观看| 亚洲综合色惰| 怎么达到女性高潮| 欧美性猛交黑人性爽| 国产亚洲精品av在线| 午夜亚洲福利在线播放| 国语自产精品视频在线第100页| 国模一区二区三区四区视频| 人妻夜夜爽99麻豆av| 日韩 亚洲 欧美在线| 欧美又色又爽又黄视频| 又爽又黄无遮挡网站| 欧美bdsm另类| 久久国产精品人妻蜜桃| 国产三级黄色录像| 久久久久久久久久成人| 最近最新免费中文字幕在线| 亚洲中文字幕一区二区三区有码在线看| 久久久国产成人免费| 亚洲精品亚洲一区二区| 久久草成人影院| 久久亚洲精品不卡| 九色国产91popny在线| 免费高清视频大片| 午夜日韩欧美国产| 女人十人毛片免费观看3o分钟| .国产精品久久| 给我免费播放毛片高清在线观看| 嫩草影视91久久| 欧美3d第一页| 久久精品国产亚洲av香蕉五月| 久久人妻av系列| 国产黄a三级三级三级人| 中文字幕av在线有码专区| 制服丝袜大香蕉在线| 黄色配什么色好看| 高清毛片免费观看视频网站| 亚洲专区中文字幕在线| 久久草成人影院| www.色视频.com| 免费观看精品视频网站| 黄色视频,在线免费观看| 欧美日韩国产亚洲二区| 深爱激情五月婷婷| 亚洲精品一卡2卡三卡4卡5卡| 精品国产亚洲在线| 国产淫片久久久久久久久 | 午夜福利视频1000在线观看| 一进一出抽搐gif免费好疼| 午夜激情欧美在线| 美女大奶头视频| 别揉我奶头 嗯啊视频| 亚洲av免费在线观看| 国产在线男女| 欧美激情在线99| 日本黄色视频三级网站网址| 一级黄片播放器| 3wmmmm亚洲av在线观看| 国产高清三级在线| 亚洲狠狠婷婷综合久久图片| 国产毛片a区久久久久| 桃红色精品国产亚洲av| 日韩免费av在线播放| 久久99热6这里只有精品| 少妇丰满av| 又黄又爽又刺激的免费视频.| 久久精品综合一区二区三区| 黄色配什么色好看| 男人和女人高潮做爰伦理| 亚洲精品亚洲一区二区| 精品久久久久久成人av| 欧美精品国产亚洲| 久久久久久久久久成人| a在线观看视频网站| 亚洲专区中文字幕在线| 成人亚洲精品av一区二区| 欧美黄色片欧美黄色片| 99久久99久久久精品蜜桃| 亚洲av第一区精品v没综合| 日本成人三级电影网站| 国产麻豆成人av免费视频| 婷婷色综合大香蕉| 夜夜躁狠狠躁天天躁| a级毛片a级免费在线| 婷婷亚洲欧美| 一级作爱视频免费观看| 美女大奶头视频| 真实男女啪啪啪动态图| 国产一区二区三区在线臀色熟女| 黄色日韩在线| 国产成人av教育| 少妇丰满av| 高潮久久久久久久久久久不卡| 我要看日韩黄色一级片| 在线观看66精品国产| 亚洲成人久久爱视频| 3wmmmm亚洲av在线观看| 啦啦啦观看免费观看视频高清| 免费大片18禁| 午夜两性在线视频| 99热6这里只有精品| 久久精品夜夜夜夜夜久久蜜豆| 久久久久久久亚洲中文字幕 | 麻豆成人av在线观看| 男人和女人高潮做爰伦理| 不卡一级毛片| 1024手机看黄色片| 久久香蕉精品热| 日韩av在线大香蕉| 两性午夜刺激爽爽歪歪视频在线观看| 欧美性猛交╳xxx乱大交人| 国产探花在线观看一区二区| 国产精品一及| 色尼玛亚洲综合影院| 亚洲片人在线观看| 国产成人啪精品午夜网站| 亚洲成人中文字幕在线播放| 美女高潮的动态| 深夜a级毛片| 老司机午夜福利在线观看视频| 美女大奶头视频| 精品日产1卡2卡| av欧美777| 欧美国产日韩亚洲一区| 99久久精品热视频| 精品久久久久久久久久久久久| 亚洲激情在线av| 乱码一卡2卡4卡精品| 宅男免费午夜| 夜夜躁狠狠躁天天躁| 天堂√8在线中文| a级一级毛片免费在线观看| 极品教师在线免费播放| 在线天堂最新版资源| 99热这里只有是精品在线观看 | 麻豆一二三区av精品| 久久亚洲真实| 此物有八面人人有两片| 日韩欧美国产在线观看| 超碰av人人做人人爽久久| 日本黄色视频三级网站网址| 免费看日本二区| 国产成年人精品一区二区| 99国产精品一区二区三区| 麻豆国产av国片精品| 男人舔女人下体高潮全视频| 亚洲av中文字字幕乱码综合| 欧美xxxx黑人xx丫x性爽| av在线观看视频网站免费| 真实男女啪啪啪动态图| 嫩草影院新地址| 悠悠久久av| 亚洲avbb在线观看| 欧美性猛交黑人性爽| 久久国产精品影院| 麻豆av噜噜一区二区三区| 精品午夜福利在线看| 又黄又爽又免费观看的视频| 18禁黄网站禁片午夜丰满| 精品一区二区免费观看| 91狼人影院| 久久性视频一级片| 精品久久久久久久久亚洲 | 成人国产一区最新在线观看| 日本三级黄在线观看| 久久久精品欧美日韩精品| 亚洲 欧美 日韩 在线 免费| 久久亚洲真实| 午夜福利在线在线| 久久久久久久亚洲中文字幕 | 国产一级毛片七仙女欲春2| 精品日产1卡2卡| 欧美日韩中文字幕国产精品一区二区三区| 久久国产精品影院| 性欧美人与动物交配| 99在线人妻在线中文字幕| 国语自产精品视频在线第100页| 久久久久性生活片| 日本精品一区二区三区蜜桃| 少妇高潮的动态图| 亚洲欧美激情综合另类| 欧美黑人欧美精品刺激| 欧美不卡视频在线免费观看| aaaaa片日本免费| 亚洲自偷自拍三级| 免费无遮挡裸体视频| 日本撒尿小便嘘嘘汇集6| 欧美黑人欧美精品刺激| 激情在线观看视频在线高清| 非洲黑人性xxxx精品又粗又长| 97超视频在线观看视频| 欧美激情久久久久久爽电影| 欧美色视频一区免费| 精品人妻视频免费看| 国产伦精品一区二区三区四那| 日韩欧美精品免费久久 | 中文字幕人妻熟人妻熟丝袜美| 深爱激情五月婷婷| 国产高潮美女av| 中文在线观看免费www的网站| 观看美女的网站| 亚洲 国产 在线| 国产精品亚洲av一区麻豆| 他把我摸到了高潮在线观看| 可以在线观看毛片的网站| 一级毛片久久久久久久久女| 日日摸夜夜添夜夜添小说| 国产免费一级a男人的天堂|