• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    SmokerViT:A Transformer-Based Method for Smoker Recognition

    2023-12-12 15:50:02AliKhanSomaiyaKhanBilalHassanRizwanKhanandZhonglongZheng
    Computers Materials&Continua 2023年10期

    Ali Khan,Somaiya Khan,Bilal Hassan,Rizwan Khan and Zhonglong Zheng,?

    1College of Mathematics and Computer Science,Zhejiang Normal University,Jinhua,321004,China

    2School of Electronics Engineering,Beijing University of Posts and Telecommunications,Beijing,100876,China

    3Department of Electrical Engineering and Computer Science,Khalifa University of Science and Technology,Abu Dhabi,127788,United Arab Emirates

    4Key Laboratory of Intelligent Education of Zhejiang Province,Zhejiang Normal University,Jinhua,321004,China

    ABSTRACT Smoking has an economic and environmental impact on society due to the toxic substances it emits.Convolutional Neural Networks(CNNs)need help describing low-level features and can miss important information.Moreover,accurate smoker detection is vital with minimum false alarms.To answer the issue,the researchers of this paper have turned to a self-attention mechanism inspired by the ViT,which has displayed state-of-the-art performance in the classification task.To effectively enforce the smoking prohibition in non-smoking locations,this work presents a Vision Transformer-inspired model called SmokerViT for detecting smokers.Moreover,this research utilizes a locally curated dataset of 1120 images evenly distributed among the two classes (Smoking and NotSmoking).Further,this research performs augmentations on the smoker detection dataset to have many images with various representations to overcome the dataset size limitation.Unlike convolutional operations used in most existing works,the proposed SmokerViT model employs a self-attention mechanism in the Transformer block,making it suitable for the smoker classification problem.Besides,this work integrates the multi-layer perceptron head block in the SmokerViT model,which contains dense layers with rectified linear activation and linear kernel regularizer with L2 for the recognition task.This work presents an exhaustive analysis to prove the efficiency of the proposed SmokerViT model.The performance of the proposed SmokerViT performance is evaluated and compared with the existing methods,where it achieves an overall classification accuracy of 97.77%,with 98.21%recall and 97.35% precision,outperforming the state-of-the-art deep learning models,including convolutional neural networks(CNNs)and other vision transformer-based models.

    KEYWORDS Smoker recognition;SmokerViT;deep learning;transformer for vision

    1 Introduction

    The smoking epidemic is one of the world’s significant public health threats,killing more than 8 million people yearly,including 1.2 million from passive smoking.In 2020 statistics[1],22.3%of the world’s population smoke,and 80%of those 1.3 billion smokers worldwide are in low-and mediumincome countries.According to a report about different causes of death worldwide,smoking is the second most significant risk factor for death [2].Smoking monitoring and preventive policies are included as actions that should be implemented in the World Health Organization(WHO)framework convention on smoking control[3].Therefore,detecting smokers in no-smoking areas is essential for effective surveillance.

    Traditional surveillance methods for smoker detection are inefficient and affected by various factors,thus limiting the development of intelligent surveillance[4,5].Researchers have continuously applied different methods to surveillance systems to answer these problems and benefit from artificial intelligence technology[6,7].Deep learning is the state-of-the-art(SOTA)artificial intelligence method that has become integral to computer vision [8].Compared to traditional image processing and machine learning methods,deep learning does not require complex image pre-processing.The Deep Neural Network (DNN),which employs deep learning techniques [9],significantly improves object detection efficiency by automatically learning the features from raw data.Convolutional Neural Networks(CNNs),a subset of DNNs,have been widely utilized to classify and cluster images based on similarity and recognize objects in scenes.CNNs have encouraged the exponential rise in deep learning as it enables significant advancements in many exciting applications,including surveillance[10,11],medical diagnosis[12],self-driving cars[13],etc.

    Since the first CNN model AlexNet [14],resulted in faster training time efficiency,new CNN models are proposed with improved accuracy with fewer parameters.Early models,such as the Visual Geometry Group (VGG) [15],had many neurons and parameters,which may result in overfitting and involve enormous computational resources.With the application of residual blocks,the training efficiency of CNN models was improved with some widely used SOTA models such as ResNet[16],Inception[17],and DenseNet[18].All the previously proposed CNN methods employed in different applications showed that accuracy is critical for applying deep learning in computer vision.

    CNN models have become a vital tool in computer vision-based surveillance applications.Convolution layers were previously widely used as the fundamental building block;however,current trends of adding attention processes have prompted researchers to rethink this.In addition to assisting CNNs with long-range dependencies,attention may replace convolutions to provide SOTA performance on computer vision tasks [19].Recently,researchers have examined using self-attention in vision-based tasks because of its potential for word-dependency learning abilities[20].Self-attention helps to learn complex relations between neighbours and their further neighbours,which may help with the binary classification problem.

    CNNs need help in describing low-level features and can miss important information.Moreover,accurate detection is vital with minimum false alarms.Considering the need for Artificial intelligence(AI) based surveillance mechanism for smoker recognition in no-smoking indoor and outdoor environments,this study focuses on the interpretation of self-attention and multilayer perceptron(MLP)head for a better understanding of the employed deep learning method.To accurately recognize smokers,this work introduces the method SmokerViT for smoker recognition in smart city indoor and outdoor environments where the Transformer component learns the convolution-like features.In SmokerViT,the patch extractor is 16 × 16 convolution with stride 16.Its output is then multiplied by learnt weights to form q,k,and v embeddings of the self-attention layer.Moreover,the MLP residual block is a linear layer in SmokerViT that raises the embedding dimension by a factor of 4,adds non-linearity then lowers it back to the original.Further linear kernelL2 classifier is used for classification.The proposed SmokerViT model uses these threefold attributes,resulting in a more robust recognition system for Smoking and NotSmoking images.The novelty of this research is to develop a deep learning-based SmokerViT model for Smoker recognition with higher accuracy.The main contributions of this research are:

    ? This research utilizes the smoker detection dataset,which has 1120 images evenly distributed among the two classes(Smoking and NotSmoking).This research performs augmentations on the dataset to have a considerable number of images with various representations to overcome the dataset size limitation.

    ? This paper presents a novel end-to-end deep learning model called SmokerViT,which integrates transformer blocks and MLP head with a fully connected layer to learn complex features and linear kernel 2 regularizer for recognizing smokers.Moreover,SmokerViT,due to its discriminative multi-head self-attention,possesses the intrinsic capabilities to classify images irrespective of the backgrounds,image quality,and noisy artefacts.

    ? This research performs exhaustive analysis to optimize the SmokerViT model to achieve the best performance on the test dataset.It can facilitate future research as a starting point for efficient smoker recognition methods.

    ? The performance of the proposed SmokerViT model is compared with different deep-learning models on the smoker detection dataset.This work uses several evaluation metrics to assess the performance of the SmokerViT model,where it outperforms the existing state-of-the-art solutions in classification accuracy.

    The research paper is organized as follows:Section 2 details the related work associated with this research,Section 3 gives the details of materials and methods adopted for solving the recognition task for the smoker detection problem,and Section 4 offers a detailed performance analysis of the proposed method and comparison with other methods and Section 5 concludes this research.

    2 Related Work

    There has been some research on various applications of surveillance using computer vision based on different proposed CNN methods.These computer vision applications include human activity recognition,pedestrian detection,traffic monitoring,face recognition,vehicle identification,fire detection,motion detection,medical imaging,etc.Authors in[21]compared state-of-the-art machine learning algorithms for insurance fraud detection.The proposed study’s decision tree algorithm performed best for the considered task.Similarly,authors in[22]proposed an improved particle swarm optimization method for data classification.Their proposed method has been tested to optimize the weight of the feed-forward neural network for fifteen datasets.Another research[23]proposed CNN based model for person head detection for counting the crowd in sports videos.Their proposed method solves the multi-scale problem,which is the object detection problem’s core issue.

    The smoker detection problem is relatively new and less explored,possibly due to the unavailability of open-access image/video datasets.Authors in [24] proposed a deep learning method based on YOLOv3-tiny named Improved YOLOv3-tiny to solve the problem of indoor low-precision smoke alarms on their local dataset.The proposed method combined the advantages of YOLOv3 and YOLOv3-tiny in terms of fewer parameters and higher accuracy for the localization task on their local smoker dataset.The proposed method considered the performance metrics of mAP for the localization task.It showed 85% mAP for improved YOLOv3-tiny compared to YOLOv3-tiny,which was 74%.However,their work limitations are the low mAP and the unavailability of the dataset.Another similar method[25]was proposed,named Eye-Smoker,YOLOv3 based transfer learning method for smoker detection on their local dataset.In the proposed method,the smoker is detected based on the cigarette and does not consider other kinds,such as e-cigarettes and smoking pipes.Their proposed method considered the localization task for smoker detection with 90%accuracy and around 94%mAP.Their limitations of work are the low accuracy and unavailability of the dataset.These object detection methods promote fast localization capabilities but lack high accuracy.

    For the classification problem,false alarms should be kept minimal.A significantly high rate of false alarms in one class and a higher number of accurate classifications in another can lead to higher prediction accuracy;however,it might lack solving the desired problem.In[26],the authors proposed a SmokingNet model based on GoogleNet for smoker detection problems on their local dataset.Their work focused on evaluating the performance of smoking and not-smoking image classification with different performance metrics.In their proposed method,the smoking image characteristics are optimized based on the GoogleNet,and the feature extraction ability is enhanced using kernels of non-square convolution.The proposed method achieved 90% accuracy,90% precision and recall,and 90% F1 measure.Their work limitations are using very basic GoogleNet as a base model and the unavailability of the dataset.In previously published work[27],the research proposed Inception-ResNet-V2-based transfer learning,where the pre-trained model was used as a backbone network for the smoker detection problem on the local smoker detection dataset.In the proposed method,the Inception-ResNet-V2 model is used,which is trained on the ImageNet dataset,the weights of the pre-trained Inception-ResNet-V2 are frozen,and new fully connected layers are added with ReLU and sigmoid activation functions.The fully connected layers learn the specific features of the task of smoker detection.The proposed method fed the complete image with an input size of 224×224 to the network.The neural network extracted the features based on the previously learned generic features trained on the ImageNet dataset.The proposed solution has a training accuracy of 95.65%and 96.87%testing accuracy with a recall of 97.32%and precision of 96.46%,discriminating the images of the Smoking and NotSmoking classes.However,the proposed work had high accuracies;still,it lacked training the model from scratch and better learn the low-level features.

    To solve the parallel processing of words by using self-attention in Recurrent Neural Network(RNN)models,a network called Transformer based on attention mechanism and removes recurrence and convolutions was proposed[28],which accomplished great success in natural language processing(NLP).After its success in NLP,an image classification model,Vision Transformer (ViT) [29],was introduced in computer vision,disrupting the traditional CNN model with its competitive performance on large-scale image datasets.With the development of transformers for computer vision in 2021,there has been some research for computer vision applications using vision-based transformers[30–32].Transformers have seen much growth in image classification tasks with accuracy similar to if not more than,CNN models.In [33],the authors proposed a multi-instance vision transformer named MITformer for remote sensing scene classification.In their proposed method,the local feature response was highlighted for the remote sensing scenes.Attention-based MLP was inserted at the end of each encoder to enhance these features.Another work [34] proposed a hybrid CNN and ViT method,CTNet,to classify high-resolution remote sensing (HRRS) images.The proposed method has two modules,T-stream(stream for ViT)and C-stream(stream for CNN).In the T-stream,the flattened patches of the image are sent into the pre-trained ViT for semantic features in HRRS images.At the same time,C-stream is used to extract the local features.Ma et al.[35]proposed a homo-heterogeneous transformer learning (HTTL) for remote sensing scene classification.In the proposed HTTL,a patch generation module is used to design homo-and heterogeneous patches.The feature learning module extracts the feature information of global and local areas.A fusion submodule and metric learning-based classification module are used for the scene classification.

    In other computer vision applications,authors [36] proposed detecting rain and road surface conditions using vision transformers.In their proposed method,a spatial self-attention network is proposed to analyze the relationship between the detection results of adjacent images as a sequenceto-sequence detection task.Dong et al.[37]proposed ViT based representation learning method for polarimetric synthetic aperture radar (PolSAR) image classification.In the proposed method,the ViT learned the global features of the PolSAR images,which improves the classification efficiency.In[38],the authors proposed a multilabel vision transformer ForestViT for multilabel classification of satellite images of deforestation,which adopts a self-attention mechanism,replacing the convolution operations.Wang et al.[39] proposed a double output vision transformer (DOViT) for air quality classification.The tokens are processed with multilabel self-attention (MSA) to extract features for higher accuracy.Authors[40]proposed Transformer based LPViT for classifying and detecting defects in printed circuit boards(PCBs).The proposed method used labels for better model strategy and mask patch prediction to ensure the relationship of different patch extractions.

    However,deep learning-based algorithms were formerly thought of as a black box,and there have been issues with their interpretability for a long time[41–43].CNNs face problems describing the lowlevel features outside the actual area of interest[44].Taking advantage of the context information for feature extraction is not beneficial.This work exploits self-attention blocks and MLP head,which are building blocks of the proposed method for visually interpreting Smoking and NotSmoking images.In the proposed method SmokerViT,which is based on Transformer and MLP head,the smoker recognition problem is considered.It achieves better prediction accuracy without convolutions than the previously proposed CNN methods.

    3 Materials and Methods

    This section details the proposed SmokerViT and the image dataset used for Smoking and NotSmoking classes for recognition problems.The following subsections explain the methodology of this work.

    3.1 Dataset Acquisition

    The dataset for this study is the smoker detection dataset published online as open access,which has different images of people smoking and not smoking indoors and outdoors.To the best of our knowledge,there is no other open-access dataset related to the problem;the smoker detection dataset facilitates future work in proposing new methods.The dataset can be accessed from[27].

    3.2 Dataset Distribution

    The smoker detection problem is considered binary,with two classes named,Smoking and NotSmoking.The NotSmoking class images are labeled 0,while the Smoking images have a class label 1.The smoker detection dataset is balanced and has 1120 images,with 560 images each in the Smoking and NotSmoking classes.This research splits the dataset into training and testing with a ratio of 80:20 with equal distribution from both classes.The training data is further divided into training and validation,with 716 images belonging to training samples and 180 for validation.

    3.3 Proposed Method

    Smoker detection in no-smoking areas is a difficult task where many factors influence the development of an AI-based surveillance system.The smoker recognition problem is solved by using a smoker detection dataset.Higher detection accuracy needs a large dataset for training,although applying deep learning models significantly improves the results.If the dataset is too small,the model is at risk of over-fitting,which means it cannot generalize effectively and will result in poor performance on a new dataset.Therefore,to train a small dataset for deep learning,this work performed data augmentations to have multiple training samples to overcome the dataset size limitation.This research implemented various augmentation processes on the training dataset,as given in Table 1.This work performed various augmentations such as resizing,scaling,flipping,shifting,etc.,as illustrated in Fig.1.All the images in the dataset are resized to a uniform resolution of 224 × 224.After that,augmentations are applied,such as vertical and horizontal shift by a factor of 0.2,50°rotation,zoom by a 0.2 factor,shear transformation,and horizontal flip by 0.2 factor.Some sample augmentation images are depicted in Fig.1.

    Figure 1:Sample data augmentations

    Table 1:Data augmentation

    To efficiently execute the recognition task,this research proposes SmokerViT inspired by Vision Transformer [29].CNNs have been pivotal in solving the problems of computer vision-based applications.In CNN,the pixels of the image are interdependent,and instead of all pixels features being trained on,only extracted features from the image patches using filters are being used for training.However,if complete data of images are used for training,the chances of obtaining the best performance become higher,which is the main work of the Transformer for vision-based applications.In the proposed SmokerViT,the work first converted the image into patches of size 16 × 16.In Transformer,the patches should be of the size that gives the equal rows and columns of the patches in the image.The image size of 224×224 and 16×16 patch size will give 14×14,196 patches per image.After the conversion into patches,it is passed to the Transformer encoder for processing.After that,the output is passed into the Multi-Layer Perceptron(MLP)head,which in the proposed SmokerViT consists of flatten,dense layer with ReLU function and a classification layer with kernel regularizerL2 to output the prediction depicted in Fig.2.

    Figure 2:Working mechanism of SmokerViT

    In SmokerViT,Vision Transformer takes the series of patches of images as input and predicts the class labels for the input image.The transformer differs from traditional CNNs,which do computations using pixel arrays.The Transformer divides the image into patches of fixed size.Then it inputs these patches into a linear projection of flattened patches embedding layer to produce vectors often known as tokens.These tokens precede a series of tokens.Additionally,the location data is provided by the position embedding.The Transformer encoder will receive these tokens as embedded patches and the location data.The Transformer encoder has the same number of outputs as inputs.The output corresponding to the class is then entered into the MLP head to output the prediction and classification.The architecture of SmokerViT is illustrated in Fig.3.

    Figure 3:Architecture of SmokerViT

    To consider the operation of SmokerViT in detail;first,the input imageXwith dimensionh×w×cis divided into several patches ofXpasn×(p2.c),wherehandwrepresent the image resolution of input while(p,p)represents the image patch resolution,cdenotes the number of channels,andrepresents the number of image patches,and this is the input sequence length for the model.These patches are then passed through a linear projection and mapped to theddimension to get the output referred to as patch embedding.The position embeddingEposis added to the patch embeddingEto keep the position information of the input.It is expressed asEpos∈R(n+1)×d,which joins the[class]token=Xclass.Its form at the output of Transformer encoderworks as image representationY.The Transformer encoder contains multi-head self-attention (MSA),layer normalization (LN) and MLP block.

    MSA: This layer linearly integrates the attention output.The encoder receives a sequence of embedding to process,which undergoes three different linear transformations to output the three vectors queryq,keyk,and valuev.The attention output for each embedding is calculated by the dot product of these three vectors.Self-attention is calculated independently and repeatedly in parallel.As a result,it is known as multi-head attention.The attention measures how strongly the patches are connected,subsequently assisting in prediction.The MSA is calculated by the equation given below:

    wheredis the dimension ofk,andwiis the learnable weights.

    LN: Layer normalization balances the mean-variance of each input neuron layer,making it converge faster.Layer normalization is added before each block,as it has no prior image dependencies,so it enhances the performance and decreases the execution time.

    MLP:The MLP in the Transformer encoder consists of two layers with GeLU.

    MLP head:After the Transformer encoder,the output is inserted into the newly added MLP head for the classification of Smoking and NotSmoking images,which consists of flatten layer to flatten the encoder output,dense layers with ReLU activation and linear kernelL2 regularizer as a classifier.

    3.3.1 Activation Function

    The activation function optimizes the processes and learns complex features specific to the task.The proposed method considers ReLU (R) activation function.Ris a piecewise linear function that outputs the input directly if it is positive;otherwise outputs zero and is given by:

    3.3.2 Optimizer

    This work considers RMSProp (Root Mean Squared Propagation) optimizer for the proposed SmokerViT method.RMSprop applies the exponential moving average of the squared gradients to adjust the learning rate.RMSprop only accumulates gradients in a specific fix window instead of letting all the gradients accumulate for momentum.The equation for RMSprop is as follows:

    whereηshows the learning rate,?represents the small term preventing division by zero,E[g2]is the past squared gradients RMSprop running average,andgtis the gradient function.

    3.3.3 Loss Function

    The binary cross entropy loss function is often used for binary classification.It helps to evaluate model accuracy by calculating the prediction probability.Following is the equation for the binary cross entropy loss function:

    wherezrepresents the label,i.e.,1 denotes the Smoking class and 0 denotes NotSmoking class,and ?zis the predicted probability ofz.

    3.3.4 Linear Kernel L2as a Classifier

    This work used the linear kernelL2 algorithm because it can help solve problems with multicollinearity (highly correlated independent variables) by limiting the coefficient and maintaining all the variables.Linear kernel,basic kernel,is the best in case of many features and given byf(whereX,Xjis the data to classify.Linear kernelL2 predicts based on the mean of data to avoid overfitting,unlikeL1,which takes the median of data for estimation.L2 adds the penalty to the cost function as the squared value of the weights and learns complex patterns.L2 is computationally efficient,and predictions are more accurate when the output is a function of all input variables.L2 regularization is calculated by:

    wherewiis the weight andλrepresents the regularization parameter.Ifλis 0,this acts as Ordinary Least Square(OLS),where it will make the weight coefficient 0 and result in underfitting,while ifλis very large,it will increase the weight and result in underfitting.

    4 Performance Evaluation

    The performance of the proposed SmokerViT for smoker recognition is evaluated and compared with other methods on the smoker detection dataset.The system configurations for simulation are i7-11800H,16 GB DDR4,NVIDIA RTX3060 6 GB,and the simulation setup is Anaconda Python 3.8 with Tensorflow 2.6 and Keras 2.3 libraries.The proposed SmokerViT is tested with various hyperparameters values for the best results.Table 2 depicts the hyper-parameters for the simulations.

    Table 2:Simulation parameters

    4.1 Evaluation Metrics

    This section presents the evaluation metrics for analyzing the performance of the proposed SmokerViT method.This work evaluated the methods on the following metrics:

    TheTnis the true negative,accurately classified as NotSmoking images,whileTpis the true positive,accurately classified as Smoking images by the proposed solution.Fnis the false negative where the Smoking image is categorized as NotSmoking,andFpis the false positive where NotSmoking images are labelled as Smoking.Precisionis the percentage of correctly positive outcomes to those the classifier predicted correctly,whereas the ratio of correctly positive results to all the relevant samples that should be positive is known as theRecallorSensitivityof the proposed method.The ratio of correct negative predictions to the results that the classifier predicts as negative is known asSpecificityor true negative rate.FDris the total number of false positive classifications to the total positive classifications.The F1 score is the harmonic mean of precision and recall,which shows how the classifier predicts correctly.FPris the ratio of negatives falsely categorized as positives and the total number of actual negatives,whileFNris the ratio of positives being falsely classified as negatives and the total number of actual positives.Eris the error rate,the ratio of all the incorrect predictions to the total number of test samples.

    4.2 Attention Maps of Learned Features

    This subsection interprets the visualization of the proposed method to understand the smoker recognition mechanism better.This research visualized the attention maps of some sample images predicted for the smoker recognition tasks from the self-attention block,illustrated in Fig.4.Selfattention is the main reason the Transformer integrates data across the complete image,including in the base layers.The attention maps show how well the method utilizes this capacity for the considered task.Some attention heads already focus on the desired representations on most images in the initial layers,demonstrating that the proposed method employs the capability to integrate information globally.The attention to the desired features increases with the model depth,and it becomes clearer what features the model pays attention to for the desired task.The original input images are converted into pseudocolour images to highlight the attention mask applied to the input image.It can be seen from the sample of images considered for both the Smoking and NotSmoking classes that the brighter part represents the attention mapping of the proposed method.Globally this research discovers that the method pays attention to image areas that are significant for classification in terms of semantics.From the maps,it can be noted that the attention of the method is on the cigarettes and their smoke for the detection of smokers.Similarly,the absence of cigarettes and smoke around the person’s mouth or hand is predicted to be a NotSmoking image.

    Figure 4:Sample images of smoking and NotSmoking(a,d)original and(b,c)attention maps

    4.3 Performance Analysis of SmokerViT

    This section presents the performance analysis of the proposed SmokerViT,a method based on the Transformer and MLP head.This work was analyzed by using different regularizers as a classifier to prove the effectiveness of linear kernel regularizer(L2)over Gaussian kernels.Moreover,this work proves the efficiency of the proposed method with and without the proposed MLP head and with and without data augmentation with simulation settings presented in Table 2.

    Table 3 presents the performance of SmokerViT in terms of prediction accuracy using different kernels.The result shows that using linear kernelL2 SmokerViT obtained the best result with 97.77%overall prediction accuracy,while the Gaussian kernel displayed overall prediction accuracy of 94.64%.It can be noted that linear kernelL2 showed the best accuracy for the Smoking and NotSmoking classes with 98.21%and 97.32%,respectively,whereas the Gaussian kernel showed 93.75%Smoking and 95.54%accuracy for NotSmoking class.

    Table 3:Performance of SmokerViT using different kernels

    This work considered the best result obtained on linear kernelL2 for the SmokerViT.This research performed further analysis by removing the MLP head and replaced with a sigmoid as a classifier with a single output to demonstrate the efficiency of the proposed model.Table 4 presents the performance of the proposed method with and without MLP head and augmentation and without augmentation block.It can be observed that the proposed model with MLP head and augmentation has the best results,whereas without augmentation performed worst for both the classes and overall prediction accuracy.The performance of SmokerVit with the proposed MLP head is improved using augmentation to 97.77% from 95.54% without the MLP head.This is because the augmentation significantly increased the dataset size with various representations,which helped predict unseen Smoking and NotSmoking images in the test dataset.While without augmentation,and proposed MLP head has significantly low performance with 88.39%overall prediction accuracy.

    Table 4:Performance of SmokerViT with and without augmentation and MLP head

    Moreover,this work performed simulations to show the effectiveness of using a Transformer as the main network over other SOTA models with anL2 kernel classifier.This work used ResNet,Inception-ResNet-V2 model to show the efficiency of using Transformer over these models.Table 5 shows that the best results are achieved using Transformer with 97.77% accuracy,while the second best results are achieved using Inception-ResNet-V2 with 96.43%accuracy,followed by InceptionV3 with 87.05% accuracy.ResNet performed worse with 85.71% accuracy.It can be observed from the table that SmokerViT has more parameters than the other models;however,the focus of this study is the higher accuracy.

    Table 5:Performance comparison of using different models for feature extraction

    The time complexity of the proposed SmokerViT model can be expressed asO(N2L),whereNrepresents the number of patches in the input image,andLrepresents the number of self-attention layers in the transformer.

    TheO(N2L)notation arises from each self-attention layer having a quadratic complexity ofO(N2),as it involves computing pairwise dot products between all pairs of patches.Since the SmokerViT model hasLself-attention layers,the total time complexity isO(N2L).

    It is worth noting that the SmokerViT model also has additional computational costs associated with the feedforward network and positional embeddings,but these are typically negligible compared to the self-attention computation.

    The loss performance curves of the proposed SmokerViT in terms of training loss and validation loss are depicted in Fig.5.The training phase of the proposed SmokerViT is carried out through 50 epochs.From the result,it can be noted that the training loss started at 76.24%and achieved less than 19.56%loss at the 10th epoch.After 10 epochs,the loss curve remained steady till the 50 epochs,with a final loss of 8.28%.Similarly,the validation loss started at 30.82% and reached 9.62% at the 10th epoch.After that,the loss curve remained steady till 50 epochs,with a final loss of 5.39%.

    Figure 5:Loss performance of the proposed SmokerViT

    The confusion matrix depicts the predictive analysis of Smoking and NotSmoking image classification.It can be seen from the confusion matrix in Fig.6 that the proposed SmokerViT displayed a prediction accuracy of 97.77% and 2.23% error rate with 109Tnand 110Tpwith 3Fpand 2Fn,respectively.Table 6 shows the performance of the proposed SmokerViT on individual classes.The proposed method achieved 98.21% prediction accuracy,97.35% precision,98.21% recall,and 97.78% F1 score for the Smoking class.Whereas for the NotSmoking class,the proposed method displayed 97.32%prediction accuracy,98.20%precision,97.32%recall,and 97.76%F1.The proposed SmokerViT displayed the overall performance with 97.77%prediction accuracy,98.21%recall,97.35%precision,and 97.78% F1 measure for classifying Smoking and NotSmoking images of the smoker detection dataset.

    Figure 6:Confusion matrix of SmokerViT

    Table 6:Performance of SmokerViT

    Receiver Operating Characteristic(ROC)curve,shown in Fig.7a,is another graphical representation for assessing the performance that shows the proposed method’s ability to predict classification with varying prediction thresholds.The ROC curve is plotted by considering Recall(TPr)on the y-axis againstFPron the x-axis.The Area under the Curve(AUC)depicts how well the method differentiates between the classes.The AUC of 0.9948 by the SmokerViT means that it has a 99.48% chance of accurately classifying the Smoking and NotSmoking classes.This work also analyzed the proposed SmokerViT based on the Precision-Recall (PR) curve,which depicts how well it performed for classifying Smoking images because,unlike ROC,the PR curve does not considerTnfor performance evaluation.PR curve also depicts whenever the class distribution has variation,unlike ROC,which shows no change.It can be noted from Fig.7b that the curve is near the top right corner,showing that the SmokerViT performed well in classifying the Smoking class.SmokerViT achieved 99.51%average precision(AP)for the Smoking and NotSmoking classification.

    Figure 7:(a)Receiver operating characteristic curve with AUC and(b)precision-recall curve with AP

    For the smoker recognition problem,false negatives should be minimal.From Fig.8b,it can be seen that the false negative occurred when the background of the image is similar to the person in the image.In computer vision,spatial resolution is crucial,which has led to the inaccurate classification of the Smoking image as NotSmoking.Better-quality images let the model generalize more accurately.In addition,the neural network had trouble distinguishing between cigarette and the background pixels in the images when the background was blurry.The lack of a considerable number of varying images in the training set might also account for the false alarms.Another possible reason is that some photos in the test set were new to the model and lacked representation of comparable images in the training set.The model performed poorly in generalizing the novel scenarios.

    Figure 8:(a)False positives and(b)false negatives

    Subsequently,a similar problem was seen with false positives.Smoker recognition relies heavily on the accuracy and practicality of the classifier,both of which are affected by the number of false positives.Images of false positives are shown in Fig.8a.The scarcity of diversity in the training set and the variety of datasets may result in the incorrect classification of some NotSmoking images as Smoking.It can also be observed image with the background as a cloud was misclassified as Smoking;moreover,a similar hand gesture to the smoking was also labeled as Smoking.

    4.4 Comparative Study with Other Methods

    For validating the effectiveness of the proposed SmokerViT,this research compared the performance with other methods,both CNN and Transformer based models such as EfficientNetV2[45],ResNest [46],MobileNetV3 [47],ResNetD [48],ViT [29],Levit [49],Davit [50] and Coatlite[51] on the smoker detection dataset.The hyperparameters are listed in Table 7.Table 8 presents the comparative analysis of these methods.It can be observed that SmokerViT displayed superiority over other considered methods in classifying the Smoking and NotSmoking classes.All the methods considered for comparison were used as pre-trained models using transfer learning and added the classification layer with a sigmoid activation function.After SmokerViT,ViT performed better among all the other considered methods for classification tasks on the local dataset for the smoker detection problem,as explained in Table 6.ViT achieved 96.43% accuracy,96.43% sensitivity,and 96.43%specificity,followed by Levit with 94.64%accuracy,Coatlite with 91.07%accuracy,Davit with 90.18%,ResNetD with 88.40%accuracy,ResNest with 86.61%accuracy,EfficientV2 with 85.27%and at last MobileNetV3 with 82.14%accuracy.It can be observed that ResNetD and Davit outperformed ViT in terms of sensitivity at 97.32% compared to ViT at 96.43%.However,MobileNetV3 performed worse in terms of specificity than other models due to the significantly large number of false positives and considerably lower number of false negatives.The proposed method performed best among other considered methods in all evaluation metrics.MobileNetV3 has the lowest accuracy of 82.14%compared to other CNN methods for analyzing the unique smoker classification problem.Fig.9 shows the performance comparison of all the methods on individual classes.SmokerViT achieves the best results on both classes,followed by ResNetD and Davit for Smoking class with 97.32%accuracy while ViT for NotSmoking class with 96.63%accuracy.

    Figure 9:Performance comparison in terms of classes of SmokerViT and other methods

    Table 7:Hyperparameters of the models

    Table 8:Comparative analysis of SmokerViT with other methods on smoker detection dataset

    4.5 Comparative Study with Previous Work

    This work compares the performance of the proposed SmokerViT with our previous work[27].Table 9 shows SmokerViT has a better overall prediction accuracy of 97.77%compared to 96.87%by our previously proposed method InceptionResNetV2.It can be noted that SmokerViT has improved the performance of smoker recognition for both classes,where it displayed 98.21% and 97.32%accuracies in discriminating the Smoking and NotSmoking images,respectively.InceptionResNetV2 achieved 97.32%and 96.43%accuracies for the Smoking and NotSmoking classes.Fig.10 compares the two methods in terms of precision,recall,and F1 measure.It can be observed that SmokerVit performed better in all the considered performance metrics,which is because the self-attention mechanism focuses on the entire image,unlike convolutions in CNN that focus on the interpretation of the high-level features rather than low-level features in classifying Smoking and NotSmoking images.

    Figure 10:Performance of SmokerViT and InceptionResNetV2

    Table 9:Comparative analysis of SmokerViT with previous work on smoker detection dataset

    In this research work,SmokerViT displayed self-attention capability and MLP Head to recognize Smoking and NotSmoking images.The results show that better performance is achieved by SmokerViT compared to the SOTA models,implying that the self-attention mechanism and MLP Head architecture may be more suitable than CNN for the Smoker recognition problem.In contrast to transformers,which can compute the attention of any patch,regardless of its distance,a CNN needs to perform additional convolutions to increase the receptive field to determine the relationship between any neighboring pixels,resulting in difficulty in possessing the ability to perform longrange computation.In SmokerViT,the patch embedding component is used to learn convolution-like features,whereas self-attention is used to learn important features and ignore the noisy ones.Results show that the SmokerViT performed better than CNN and Transformer based models,validating the superiority of using both the self-attention mechanism and MLP Head.

    While looking at the results,it can be observed that the SmokerViT has performed better in both the Smoking and NotSMoking classes.However,CNNs models were better at predicting Smoking images while poorly classifying NotSmoking class compared to Transformer based models,which performed well in classifying both the Smoking and NotSmoking classification.While SmokerViT was equally good in classifying both classes indicating that SmokerViT is more robust than using CNN or transformer-based models in dealing with balanced datasets.

    Many researchers used CNN for the smoker detection problem,but there is not much work on this problem.This is the first time using transformers and MLP Head with Linear kernelL2 classifier for the smoker recognition task.Additionally,previous research results were compared with this proposed work,as depicted in Table 9.It can be observed that SmokerViT outperformed the Inception-ResNet-V2 model for Smoking and NotSmoking image classification using the same dataset.Moreover,the high accuracy displayed by the proposed method can help an AI-based smoker detection system and save time and human resources simultaneously.This research can benefit researchers to improve further the methodology for image segmentation to detect cigarette smoker detection problems.

    5 Conclusion

    This research proposed a transformer-based smoker recognition method.For effective surveillance of the no-smoking areas,this research proposed SmokerViT based on the self-attention mechanism instead of CNN.The dataset for this work has two classes with 560 images each for the Smoking and Notsmoking classes.Further,this work performed augmentations on the smoker detection dataset to have many images with various representations to overcome the dataset size limitation.The proposed SmokerViT is inspired by Vision Transformer and adding our own MLP head block,which has a dense layer with ReLU activation function and linear kernelL2 regularizer as a classifier.SmokerViT extracted features through long-range dependency compared to CNN models,which took advantage of useful global information.Ablations were performed on the proposed SmokerViT to prove the efficiency of the MLP head block and data augmentation.The SmokerViT performance was evaluated and compared with the previously proposed CNN model for the smoker detection problem and other Transformer and CNN-based methods.The SmokerViT achieved a 0.93%higher accuracy of 97.77%,with 0.92%better recall of 98.21%and 0.92%better precision of 97.35%compared to the previous proposed Inception-ResNet-V2 based transfer learning method.Moreover,the results showed that SmokerViT achieved competitive performance compared to other models with considerably higher values of the evaluation metrics.

    For future works,several issues that were not addressed in this research need to be observed.The effect of dataset size on the training performance of the method and data augmentation by various complex models for ensuring further variances of the image representations can be considered for future study.Moreover,the weak point of this research is the higher number of parameters and high execution time.In future work,this point would be considered for designing the model,which is lightweight and,at the same time,yield higher accuracy.Moreover,the hybrid method of convolutions and transformer might help the smoker recognition system to perform better,considering the shortcomings of the proposed method.

    Acknowledgement:Authors are thankful to the reviewers for their valuable comments.

    Funding Statement:The authors received no specific funding for this study.

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:A.K.;data collection:A.K.;analysis and interpretation of results:A.K.,S.K.,B.H.and R.K.;draft manuscript preparation:A.K.,S.K.,B.H.and R.K.;writing—review and editing,A.K.,S.K.,B.H.,R.K.and Z.Z.;funding acquisition,A.K.and Z.Z.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The dataset considered in this study can be accessed from https://data.mendeley.com/datasets/j45dj8bgfc/1.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    狂野欧美激情性xxxx在线观看| 久久久a久久爽久久v久久| 日本三级黄在线观看| 国产午夜精品一二区理论片| 午夜福利网站1000一区二区三区| av又黄又爽大尺度在线免费看| 性插视频无遮挡在线免费观看| 亚洲精品456在线播放app| 国产伦精品一区二区三区四那| 欧美xxⅹ黑人| 国产亚洲91精品色在线| 毛片女人毛片| 一级毛片aaaaaa免费看小| 亚洲电影在线观看av| 国产午夜精品久久久久久一区二区三区| 亚洲国产欧美人成| 亚洲av免费高清在线观看| 国产乱人偷精品视频| 国产毛片a区久久久久| 街头女战士在线观看网站| av一本久久久久| 成年人午夜在线观看视频| 新久久久久国产一级毛片| 国产精品.久久久| 久久国产乱子免费精品| 欧美日韩视频高清一区二区三区二| 一级黄片播放器| 成人欧美大片| 汤姆久久久久久久影院中文字幕| 久久综合国产亚洲精品| 国产精品久久久久久久久免| 国产白丝娇喘喷水9色精品| 五月伊人婷婷丁香| 最近中文字幕高清免费大全6| 性色av一级| 国产精品人妻久久久久久| 欧美精品人与动牲交sv欧美| 国产视频首页在线观看| 欧美xxxx性猛交bbbb| 成人欧美大片| 日本与韩国留学比较| 国产高清国产精品国产三级 | av在线app专区| 国产黄片美女视频| 免费电影在线观看免费观看| 久久久色成人| 亚洲精华国产精华液的使用体验| 黄片无遮挡物在线观看| 日本爱情动作片www.在线观看| 天堂俺去俺来也www色官网| 看黄色毛片网站| 亚洲精品乱码久久久久久按摩| 一级a做视频免费观看| 一本一本综合久久| 赤兔流量卡办理| 国产乱人视频| 有码 亚洲区| 欧美激情在线99| 激情五月婷婷亚洲| 18禁裸乳无遮挡动漫免费视频 | 人妻系列 视频| 成人鲁丝片一二三区免费| 97在线人人人人妻| 国产精品人妻久久久久久| .国产精品久久| 婷婷色麻豆天堂久久| 少妇人妻久久综合中文| 男女国产视频网站| 丝袜美腿在线中文| 日韩 亚洲 欧美在线| 九九爱精品视频在线观看| 赤兔流量卡办理| 精品国产露脸久久av麻豆| 国产黄频视频在线观看| 丰满人妻一区二区三区视频av| 国产综合精华液| 超碰97精品在线观看| 特级一级黄色大片| eeuss影院久久| 久久6这里有精品| 国产爽快片一区二区三区| 网址你懂的国产日韩在线| 日本wwww免费看| 国模一区二区三区四区视频| 联通29元200g的流量卡| 街头女战士在线观看网站| 观看美女的网站| 亚洲精品456在线播放app| 九草在线视频观看| 国产精品国产三级国产av玫瑰| 男插女下体视频免费在线播放| 一级爰片在线观看| 亚洲av国产av综合av卡| 国产av国产精品国产| 内地一区二区视频在线| 黄色配什么色好看| av免费观看日本| 深爱激情五月婷婷| 在线亚洲精品国产二区图片欧美 | av国产免费在线观看| 18禁在线播放成人免费| 欧美日韩亚洲高清精品| 国产精品一区二区在线观看99| 免费观看a级毛片全部| 国产黄片视频在线免费观看| 九九爱精品视频在线观看| 国产白丝娇喘喷水9色精品| 在线观看一区二区三区| 国内少妇人妻偷人精品xxx网站| 免费不卡的大黄色大毛片视频在线观看| 国产免费一区二区三区四区乱码| 国产日韩欧美亚洲二区| av天堂中文字幕网| 一级片'在线观看视频| 国产精品熟女久久久久浪| 国产高清三级在线| 边亲边吃奶的免费视频| 国产亚洲最大av| 久久精品夜色国产| 91精品一卡2卡3卡4卡| 男的添女的下面高潮视频| 国产亚洲91精品色在线| av女优亚洲男人天堂| 亚洲三级黄色毛片| 国产永久视频网站| 亚洲av一区综合| 欧美 日韩 精品 国产| 亚洲成色77777| av免费在线看不卡| 免费在线观看成人毛片| 久久女婷五月综合色啪小说 | 久久久久久久久久久丰满| 日本免费在线观看一区| 久久综合国产亚洲精品| 久久久a久久爽久久v久久| 日韩欧美 国产精品| 亚洲av电影在线观看一区二区三区 | 精品一区二区三卡| 成人黄色视频免费在线看| 久久韩国三级中文字幕| 成人欧美大片| 精品国产三级普通话版| 熟女人妻精品中文字幕| 女人久久www免费人成看片| 美女国产视频在线观看| 免费在线观看成人毛片| 99视频精品全部免费 在线| 国产亚洲午夜精品一区二区久久 | 久久久成人免费电影| 热99国产精品久久久久久7| 人妻 亚洲 视频| 国产一区二区三区av在线| 一区二区三区免费毛片| 国产探花在线观看一区二区| av国产免费在线观看| 亚洲综合精品二区| av在线亚洲专区| 亚洲av成人精品一二三区| 亚洲最大成人手机在线| 日韩,欧美,国产一区二区三区| 亚洲av男天堂| 韩国高清视频一区二区三区| 插逼视频在线观看| 有码 亚洲区| 亚洲欧美清纯卡通| 精品99又大又爽又粗少妇毛片| 黄色欧美视频在线观看| 久久久久国产网址| 日韩制服骚丝袜av| 国内精品宾馆在线| 99热这里只有是精品50| 1000部很黄的大片| 亚洲精品国产成人久久av| 大片免费播放器 马上看| 人妻系列 视频| 亚洲丝袜综合中文字幕| 你懂的网址亚洲精品在线观看| 亚洲av成人精品一二三区| 亚洲,一卡二卡三卡| 成年免费大片在线观看| 午夜视频国产福利| 1000部很黄的大片| 2018国产大陆天天弄谢| 日韩亚洲欧美综合| tube8黄色片| 国产精品久久久久久久电影| 日韩制服骚丝袜av| 日韩一区二区视频免费看| 国产av码专区亚洲av| 亚洲自偷自拍三级| 国产片特级美女逼逼视频| 在线观看一区二区三区激情| 日本午夜av视频| 春色校园在线视频观看| 岛国毛片在线播放| 在线观看国产h片| 亚洲图色成人| 亚洲真实伦在线观看| 一区二区三区免费毛片| 日日摸夜夜添夜夜添av毛片| 各种免费的搞黄视频| 日韩一区二区三区影片| 国产免费一级a男人的天堂| 一级毛片久久久久久久久女| 男人添女人高潮全过程视频| 美女主播在线视频| 成年人午夜在线观看视频| 国产成人freesex在线| 各种免费的搞黄视频| 超碰97精品在线观看| 午夜福利在线观看免费完整高清在| 我要看日韩黄色一级片| 欧美97在线视频| 国产 一区精品| 热99国产精品久久久久久7| 一个人看视频在线观看www免费| 亚洲精品色激情综合| 男人狂女人下面高潮的视频| 国产乱人视频| 国产精品熟女久久久久浪| 伊人久久精品亚洲午夜| 噜噜噜噜噜久久久久久91| 精品亚洲乱码少妇综合久久| 国产精品国产三级专区第一集| 男男h啪啪无遮挡| 精品久久久久久久人妻蜜臀av| 国内揄拍国产精品人妻在线| 国产视频内射| 国内少妇人妻偷人精品xxx网站| 午夜福利视频精品| 91狼人影院| 亚州av有码| 国产精品久久久久久精品古装| 国产熟女欧美一区二区| 欧美成人a在线观看| 人人妻人人看人人澡| 我的女老师完整版在线观看| 欧美极品一区二区三区四区| 91午夜精品亚洲一区二区三区| 男男h啪啪无遮挡| 国产精品人妻久久久久久| 最近最新中文字幕免费大全7| 久久精品夜色国产| 永久免费av网站大全| 国产成人免费无遮挡视频| 成年女人在线观看亚洲视频 | 特大巨黑吊av在线直播| 成人毛片a级毛片在线播放| 91久久精品国产一区二区成人| 久久这里有精品视频免费| 亚洲国产av新网站| 欧美成人精品欧美一级黄| 亚洲无线观看免费| 午夜福利在线在线| 一级二级三级毛片免费看| 男人狂女人下面高潮的视频| 精品少妇黑人巨大在线播放| 视频区图区小说| 亚洲综合精品二区| 亚洲三级黄色毛片| 亚洲精品久久久久久婷婷小说| 成人漫画全彩无遮挡| 成人二区视频| 久久久久久久久久久免费av| 成人国产av品久久久| 久久99精品国语久久久| 91精品伊人久久大香线蕉| 97超碰精品成人国产| 一本色道久久久久久精品综合| 国产v大片淫在线免费观看| 午夜福利高清视频| 中文乱码字字幕精品一区二区三区| 国产精品一区二区三区四区免费观看| 欧美日韩视频高清一区二区三区二| 男女无遮挡免费网站观看| 草草在线视频免费看| a级毛色黄片| 1000部很黄的大片| 亚洲av免费高清在线观看| 一个人看的www免费观看视频| 搡老乐熟女国产| 夜夜爽夜夜爽视频| 亚洲国产精品专区欧美| 国产久久久一区二区三区| 一区二区三区精品91| 精品一区二区三区视频在线| 精品国产露脸久久av麻豆| 亚洲,一卡二卡三卡| 国产精品人妻久久久久久| 自拍欧美九色日韩亚洲蝌蚪91 | 欧美成人精品欧美一级黄| 婷婷色麻豆天堂久久| 久久久久网色| 欧美xxⅹ黑人| 亚洲最大成人手机在线| 免费观看在线日韩| 18禁裸乳无遮挡免费网站照片| 亚洲最大成人中文| 校园人妻丝袜中文字幕| 91狼人影院| 欧美另类一区| 黄色欧美视频在线观看| 丰满人妻一区二区三区视频av| 精品99又大又爽又粗少妇毛片| 黄片无遮挡物在线观看| 成人鲁丝片一二三区免费| 亚洲精品自拍成人| 在线 av 中文字幕| 制服丝袜香蕉在线| 老司机影院成人| 色网站视频免费| 女人十人毛片免费观看3o分钟| 国产伦在线观看视频一区| 一级a做视频免费观看| 国产一区有黄有色的免费视频| 亚洲怡红院男人天堂| 在线亚洲精品国产二区图片欧美 | 观看免费一级毛片| 久久久久久伊人网av| 免费黄网站久久成人精品| 街头女战士在线观看网站| 少妇 在线观看| 人妻夜夜爽99麻豆av| 性色avwww在线观看| h日本视频在线播放| 亚洲色图综合在线观看| 国产一区二区三区综合在线观看 | 超碰97精品在线观看| 国产熟女欧美一区二区| 九色成人免费人妻av| 国产老妇伦熟女老妇高清| 少妇的逼水好多| 亚洲精品成人久久久久久| 97在线视频观看| 欧美成人精品欧美一级黄| 国产高清不卡午夜福利| 亚洲内射少妇av| 欧美成人精品欧美一级黄| 国产午夜福利久久久久久| 久久影院123| 精品99又大又爽又粗少妇毛片| 人妻夜夜爽99麻豆av| 国产高清有码在线观看视频| 丝袜喷水一区| 国产极品天堂在线| 欧美日韩国产mv在线观看视频 | 内射极品少妇av片p| 亚洲精品乱久久久久久| 国产精品久久久久久精品古装| 你懂的网址亚洲精品在线观看| 青青草视频在线视频观看| 国产大屁股一区二区在线视频| 亚洲熟女精品中文字幕| 国产亚洲精品久久久com| 在线a可以看的网站| 在线观看免费高清a一片| 日本熟妇午夜| 在线观看免费高清a一片| 欧美激情久久久久久爽电影| 亚洲aⅴ乱码一区二区在线播放| 国产一区二区三区综合在线观看 | 99热这里只有精品一区| 国产精品秋霞免费鲁丝片| 欧美高清性xxxxhd video| 日韩精品有码人妻一区| 亚洲精品一区蜜桃| 99久久中文字幕三级久久日本| 新久久久久国产一级毛片| 国产成人免费观看mmmm| 亚洲在线观看片| 亚洲av免费在线观看| 国产综合懂色| 97超视频在线观看视频| 久久精品熟女亚洲av麻豆精品| 国产亚洲91精品色在线| 国产在线男女| 美女xxoo啪啪120秒动态图| 国产伦理片在线播放av一区| 久久精品久久精品一区二区三区| 三级经典国产精品| 久久久久九九精品影院| 制服丝袜香蕉在线| 水蜜桃什么品种好| 久久久午夜欧美精品| 久久人人爽人人片av| 亚洲av国产av综合av卡| 亚洲精品中文字幕在线视频 | www.av在线官网国产| 亚洲aⅴ乱码一区二区在线播放| 少妇高潮的动态图| 嫩草影院入口| 三级男女做爰猛烈吃奶摸视频| av国产久精品久网站免费入址| .国产精品久久| 性色avwww在线观看| 亚洲国产高清在线一区二区三| 亚洲最大成人手机在线| 久久久亚洲精品成人影院| 97精品久久久久久久久久精品| 亚洲av中文av极速乱| 99热全是精品| 精品久久久久久久久亚洲| 亚洲av不卡在线观看| 少妇人妻 视频| 精品99又大又爽又粗少妇毛片| 国产精品秋霞免费鲁丝片| 新久久久久国产一级毛片| 爱豆传媒免费全集在线观看| 国产伦精品一区二区三区四那| 国产精品爽爽va在线观看网站| 人妻一区二区av| 嫩草影院新地址| 成年免费大片在线观看| av国产免费在线观看| 亚洲久久久久久中文字幕| 亚洲国产av新网站| 国产免费又黄又爽又色| 欧美另类一区| 国产成人freesex在线| 91精品伊人久久大香线蕉| 精品国产一区二区三区久久久樱花 | 亚洲av一区综合| 亚洲国产日韩一区二区| 精品久久久精品久久久| 亚洲精品aⅴ在线观看| 在线免费观看不下载黄p国产| 欧美亚洲 丝袜 人妻 在线| 国产精品伦人一区二区| 一区二区av电影网| 国产成人a区在线观看| 色哟哟·www| 中文资源天堂在线| 久久韩国三级中文字幕| 欧美xxⅹ黑人| 久久精品国产亚洲网站| 久久精品国产自在天天线| 亚洲国产精品国产精品| 一区二区av电影网| 欧美成人精品欧美一级黄| 插逼视频在线观看| 午夜福利高清视频| 一级毛片我不卡| 亚洲精品一二三| 啦啦啦在线观看免费高清www| 乱码一卡2卡4卡精品| 久久精品国产自在天天线| 在线免费观看不下载黄p国产| 国产熟女欧美一区二区| 亚洲av一区综合| 亚洲av中文字字幕乱码综合| 亚洲自偷自拍三级| 极品少妇高潮喷水抽搐| 3wmmmm亚洲av在线观看| 亚洲精品久久午夜乱码| 亚洲精品第二区| 欧美精品一区二区大全| 国精品久久久久久国模美| 国产毛片在线视频| av又黄又爽大尺度在线免费看| 久久精品国产a三级三级三级| 国产午夜精品久久久久久一区二区三区| 免费大片18禁| 大片免费播放器 马上看| 久久久久国产网址| 男女国产视频网站| 久久精品熟女亚洲av麻豆精品| 中国三级夫妇交换| 国产高清国产精品国产三级 | 欧美xxxx黑人xx丫x性爽| 精品酒店卫生间| 国产av国产精品国产| 久久精品国产亚洲av涩爱| 青青草视频在线视频观看| 亚洲精品一二三| 久久久精品欧美日韩精品| 人人妻人人看人人澡| 激情五月婷婷亚洲| av专区在线播放| 亚洲欧洲日产国产| av专区在线播放| 99久久中文字幕三级久久日本| 欧美成人午夜免费资源| 一区二区三区精品91| 国产美女午夜福利| 久久久久久久久久人人人人人人| 亚洲av一区综合| 日韩一区二区视频免费看| av免费在线看不卡| 婷婷色av中文字幕| 汤姆久久久久久久影院中文字幕| 51国产日韩欧美| 男女边摸边吃奶| 国产美女午夜福利| 国产极品天堂在线| 国产精品久久久久久精品古装| 日韩中字成人| 亚洲精品国产av蜜桃| 一区二区av电影网| 丝瓜视频免费看黄片| 少妇人妻一区二区三区视频| 天堂俺去俺来也www色官网| 国产极品天堂在线| 少妇丰满av| 大香蕉97超碰在线| 欧美日韩一区二区视频在线观看视频在线 | 久久99热这里只有精品18| 免费看av在线观看网站| 午夜福利在线观看免费完整高清在| 久久99热这里只频精品6学生| 国产淫片久久久久久久久| 纵有疾风起免费观看全集完整版| 日本欧美国产在线视频| 永久免费av网站大全| 精品人妻熟女av久视频| 国内少妇人妻偷人精品xxx网站| 国产人妻一区二区三区在| 成人毛片60女人毛片免费| 一级毛片我不卡| 夜夜看夜夜爽夜夜摸| 99久久精品国产国产毛片| 天堂中文最新版在线下载 | 久久久久精品性色| 国内精品美女久久久久久| 男插女下体视频免费在线播放| 午夜福利视频1000在线观看| 99久久九九国产精品国产免费| 69人妻影院| 搞女人的毛片| 国产综合精华液| 精品少妇黑人巨大在线播放| .国产精品久久| 日韩成人av中文字幕在线观看| 亚洲人成网站在线观看播放| 一级毛片电影观看| 国产伦在线观看视频一区| 我要看日韩黄色一级片| 国产一区有黄有色的免费视频| 精品午夜福利在线看| 国产日韩欧美在线精品| 一级片'在线观看视频| 小蜜桃在线观看免费完整版高清| 久久久精品欧美日韩精品| 国产探花极品一区二区| 免费观看无遮挡的男女| 国产男女超爽视频在线观看| 最近手机中文字幕大全| 国产免费福利视频在线观看| 成年人午夜在线观看视频| 亚洲精品乱码久久久久久按摩| 成年免费大片在线观看| 五月伊人婷婷丁香| 久久精品国产鲁丝片午夜精品| 中国三级夫妇交换| 免费人成在线观看视频色| 久久精品综合一区二区三区| 久久亚洲国产成人精品v| 久久精品久久精品一区二区三区| 一个人看的www免费观看视频| 日韩免费高清中文字幕av| 大话2 男鬼变身卡| 免费黄色在线免费观看| 亚洲成人精品中文字幕电影| 国产爽快片一区二区三区| 免费电影在线观看免费观看| 免费人成在线观看视频色| 激情 狠狠 欧美| 国产欧美亚洲国产| 欧美日韩国产mv在线观看视频 | 一级毛片黄色毛片免费观看视频| 成年人午夜在线观看视频| 亚洲精品视频女| 涩涩av久久男人的天堂| 日韩成人av中文字幕在线观看| 1000部很黄的大片| 丝瓜视频免费看黄片| 听说在线观看完整版免费高清| 亚洲第一区二区三区不卡| 亚洲精品456在线播放app| 九色成人免费人妻av| 免费av不卡在线播放| freevideosex欧美| av黄色大香蕉| 天美传媒精品一区二区| 成人毛片60女人毛片免费| 高清毛片免费看| 18禁在线播放成人免费| 午夜福利网站1000一区二区三区| 久久精品国产亚洲av天美| 亚洲最大成人av| 精品熟女少妇av免费看| 欧美日韩亚洲高清精品| 国产亚洲最大av| 精品熟女少妇av免费看| 狠狠精品人妻久久久久久综合| 精品一区在线观看国产| 免费观看在线日韩| 国产精品国产三级国产专区5o| 最新中文字幕久久久久| 一级爰片在线观看| 听说在线观看完整版免费高清| 国产视频内射| 久久久午夜欧美精品| 免费黄网站久久成人精品| 麻豆国产97在线/欧美| 日韩,欧美,国产一区二区三区| 国产精品嫩草影院av在线观看| 少妇人妻一区二区三区视频| 国语对白做爰xxxⅹ性视频网站| 午夜福利在线在线| 内地一区二区视频在线| 一本色道久久久久久精品综合| 国产又色又爽无遮挡免| 国产成人a∨麻豆精品| 欧美区成人在线视频| 美女xxoo啪啪120秒动态图|