• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Intelligent Dynamic Gesture Recognition Using CNN Empowered by Edit Distance

    2021-12-15 12:48:56ShaziaSaqibAllahDittaMuhammadAdnanKhanSyedAsadRazaKazmiandHaniAlquhayz
    Computers Materials&Continua 2021年2期

    Shazia Saqib,Allah Ditta,Muhammad Adnan Khan,*,Syed Asad Raza Kazmi and Hani Alquhayz

    1Department of Computer Science, Lahore Garrison University, Lahore, 54000,Pakistan

    2Department of Information Sciences, Division of Science &Technology, University of Education, Lahore, 54000,Pakistan

    3GC University, Lahore, 54000, Pakistan

    4Department of Computer Science and Information, College of Science in Zulfi,Majmaah University, Al-Majmaah, 11952,Saudi Arabia

    Abstract: Human activity detection and recognition is a challenging task.Video surveillance can benefit greatly by advances in Internet of Things(IoT)and cloud computing.Artificial intelligence IoT (AIoT) based devices form the basis of a smart city.The research presents Intelligent dynamic gesture recognition (IDGR)using a Convolutional neural network (CNN) empowered by edit distance for video recognition.The proposed system has been evaluated using AIoT enabled devices for static and dynamic gestures of Pakistani sign language (PSL).However, the proposed methodology can work efficiently for any type of video.The proposed research concludes that deep learning and convolutional neural networks give a most appropriate solution retaining discriminative and dynamic information of the input action.The research proposes recognition of dynamic gestures using image recognition of the keyframes based on CNN extracted from the human activity.Edit distance is used to find out the label of the word to which those sets of frames belong to.The simulation results have shown that at 400 videos per human action, 100 epochs, 234 × 234 image size, the accuracy of the system is 90.79%, which is a reasonable accuracy for a relatively small dataset as compared to the previously published techniques.

    Keywords: Sign languages;keyframe;edit distance;misrate;accuracy

    1 Introduction

    Video content(a sequence of 2D frames)is globally growing exponentially every year.As a result,lots of effort has been made in the image and video recognition domain.Video classification and video captioning are two major active research areas at the moment.Video classification recognizes these videos using their content while the video captioning gives a short description of these videos using their content.Video classification is done in the spatial domain as well as in the temporal domain either separately or collectively.Convolutional neural networks (CNN) has given promising performance for analyzing image content,image recognition,detection,and retrieval.These networks can process millions of parameters and handle huge labelled datasets for learning.This has led to the testing of CNN in large scale video classification, in static images as well as in complex temporal evolution of the videos.

    Processing raw video sequences are not efficient as they have very high dimensionality depending on image dimensions and video duration.A video is a sequence of images.Most images in a video do not contain any new information.These images keep repeating, and usually after 10-15 frames approximately, a new chunk of data from the video appears and is vital for action recognition [1].This leads to the use of keyframes for action recognition.The keyframe represents valuable information in that temporal segment.The edit distance identifies the class of the video using the summarized keyframes.To recognize videos, several supervised and unsupervised techniques have been used that are based on bio-inspired sign descriptors, border(boundary) following, chain codes, polygonal approximation, Fourier descriptors, polygonal approximation, Fourier descriptors, statistical descriptors,regional descriptors, and deep learning [2].However, deep learning-based techniques have given better results than all other techniques.

    CNN can perform well if the system works with reliable datasets and GPUs.But still, many issues remain to make the system robust and practical.These are some of the problems:

    1.1 Huge Datasets Required

    All recognition systems depend on the extensive collection of videos.In many situations,a large video training set may not be available,So this puts some limitations on the use of CNN for recognition systems.We need to work with networks that can give good results with reasonably sized training data.

    1.2 Invariance

    The recognition systems must be invariant to translation rotation and scaling.While dealing with video invariances in 3D is needed.

    1.3 Handling Degradations in Training Data

    The networks should be robust to low resolution, blurring, pose variations, illumination, and occlusion [3].

    1.4 Structure of the Network

    The decisions like the number of layers,fully connected layers,dropouts,max-pooling operations can affect the efficiency of the CNN[3].

    1.5 Training and Validation Set Generation

    To determine the performance of the network, we divide our video data set to training validation and testing.

    1.6 Early Fusion

    The early fusion methods combine input features from various modalities.The fusion is done immediately on the lowest possible level, which is a pixel level.The network learns the correlation and interactions of each modality.Early fusion performs multimodal learning.It usually requires the features from different modalities to align with their semantics.It uses a single model to predict, which shows that the model is well suited for all the modalities.The early and direct connectivity to pixel data allows the network to detect local motion speed and direction [3]precisely.

    1.7 Exploding Vanishing Gradients

    The problem requires to use low learning rates with gradient descent.For a slow computer,this process will take a long time for each step.A faster GPU can overcome this delay.Another way to handle this problem is to add more hidden layers which help the network to learn more complex arbitrary functions,and in predicting future outcomes.

    The paper layout is as explained:Section 2 shows previous work done in the past in this domain,Section 3 elaborates on the experimental work based on the algorithm written in Section 3.Section 4 discusses the outcomes of the experiment, Section 6 compares the proposed system with the existing techniques and Section 7 gives a conclusion and suggests future work.

    2 Related Work

    Kanehira et al.[1] proposed Fisher’s discriminant criteria for an inner summary, inner group, and between-group variances defined on the feature representation of summary.SE De Avila et al.[2]proposed Video summarization (VSUMM) for producing static video summaries using the k-means clustering algorithm.Sebastian et al.[3] have used the mean, variance, skew, and kurtosis histogram of every block and compared it with the corresponding blocks of the next frame.Kamoji et al.[4] have analyzed the motion, block matching techniques based on diamond search, and three-step search.Gong et al.[5] has used the Sequential determinantal point process (SEQDPP) for keyframe selection based on a random permutation of video frames.

    Cahuina et al.[6] have proposed a technique of using local descriptors for semantic video summarization.They tested the method on 100 videos.Shi et al.[7] have proposed a keyframe extraction method for video copyright protection.Mahmoud et al.[8] have suggested the use of VGRAPH that used color as well as texture features.Guan et al.[9] have suggested a key point-based framework to select keyframes using local features.Asade et al.[10] suggested an algorithm to extract static video summaries using fuzzy c-means clustering.Kim et al.[11] have proposed a technique that generates panoramic images from web-based geographic information systems using data fusion, crowdsourcing, and recent advances in media processing.Danelljan et al.[12] used Discriminative correlation filters (DCF) for visual object tracking.Wang et al.[13] proposed dense trajectories to recognize an action in videos.Surf descriptors and dense optical flow were used to compare feature points for estimating homographs.This significantly improves motion-based descriptors, such as Histograms of optical flow (HOF) and Motion boundary histograms (MBH).Experimental results on four challenging action datasets (i.e., Hollywood2,HMDB51, Olympic Sports, and UCF50) give better results than other techniques.Bansal et al.[14]have used the hidden Markov model as an indispensable tool for the recognition of dynamic gestures in real-time.Bhuyan [15] has proposed gesture spotting to eliminate the effects of changes in a“Motion chain code (MCC)”.Kalman filter determines the track of each person [16].Mei et al.[17]proposed a constrained Minimum sparse reconstruction (MSR) model-based Video summarization (VS).Muhammad et al.[18] have used an effective shot segmentation method based on deep features.They have used entropy along with memorability testing their algorithm on two video datasets.Burec et al.[19]have used models inspired by human models using estimation of joint trajectories, and spatiotemporal local descriptors.

    Panda et al.[20] has used graph clustering based on random walks using a factor-based ranking.Voulodimoswe et al.[21] have used k means ++ clustering as well as temporal summaries on two dance motion datasets and got promising results.Zhang et al.[22] have used semantic side information of video to generate subshot-based summarization at significantly less computational costs testing their work on several benchmarks and got promising results.Chellappa et al.[23] has used Deep convolutional neural networks (DCNNs) for face and other objects recognition giving promising results.

    Singha et al.[24]has developed a classifier fusion based dynamic freehand gesture recognition system using a two-level speed normalization procedure based on Dynamic time warping (DTW) and Euclidean distance.Pigou et al.[25] proposed a technique that uses a simple pooling strategy using the temporal aspect of the video.Varol et al.[26] have used Long term temporal convolutions (LTC).They proved that LTC-CNN models give an increased accuracy of action recognition.Jiang et al.[27] have proposed a framework using the feature relationships and the class relationships by imposing regularization thus offering Regularized deep neural networks (rDNN) for modelling video semantics getting reasonable results on Hollywood2 and Columbia video benchmarks.Donahue et al.[28] have used recurrent convolutional architectures for image captioning, activity recognition, and video description.Simonyan et al.[29] have proposed a spatiotemporal ConvNet architecture that uses multi-frame dense optical flow on limited training data.They tested their technique on CF-101 and HMDB-51.Tran et al.[30] proposed architecture for spatiotemporal feature learning using 3 3 3 convolution kernels.It gives the best architectures for 3D ConvNets.The learned Convolutional3D (C3D) features along with a linear classifier giving 52.8% accuracy on UCF101 dataset.Thakre et al.[31] have made use of video partitioning and keyframe extraction for video analysis and content-based video retrieval.Sheena et al.[32] have proposed a method that uses the difference of histograms in consecutive frames calculates the mean and standard deviation of the difference between frames.Then using these values threshold is calculated.The experiments are conducted on the KTH action database.Ng et al.[33] have used Long Short Term Memory (LSTM) cells based Recurrent Neural Network.Lillicrap et al.[34] have worked on Deep Q-learning, a technique, based on the “deterministic policy gradient”.Redmon et al.[35] have a deep learning neural network called YOLO.Ren et al.[36] have used the region proposal algorithms to identify object locations called Region proposal network (RPN).Nam et al.[37] have worked on Convolutional neural networks (CNNs).Bertinetto et al.[38] has used Stochastic gradient descent to adjust the weights of the network, compromising the speed of the system.Feichtenhofer et al.[39] have suggested ways of fusing ConvNet towers in the spatial and the temporal domain.

    Zhu et al.[40]has proposed a full“visual tracking procedure”in videos using“Reinforcement Learning Algorithms”.Song et al.[41] have presented Title-based video summarization (TVSUM).Zaal et al.[42]have used algorithm Iterative self-organizing data analysis technique (ISODATA) to cluster frames into classes automatically.Saqib et al.[43] has proposed a method for video summarization based on entropy and mean of frames which use human activity recognition in place of the full video.Ejaz et al.[44] have combined the features of Red Green Blue (RGB) color channels, histograms, and moments to find the keyframes.The technique is adaptive as it combines current and old iterations.Jaouedi et al.[45] have used hybrid deep learning based on Gated recurrent neural networks (GRNN) model for human action recognition tested on UCF Sports, UCF101, and KTH datasets.The research analyses videos and extract related features using GMM and KF methods.The visual characteristic of each frame from the input video is used along with recurrent neural networks model based on the gated recurrent unit.The research analyses and extracts all features in all frames of video.The research applies to a wide range of applications.

    3 Proposed Solution

    The proposed technique is based on the Convolution neural network(CNN).Fig.1 shows a transition from pixel to actual object recognition using 3 hidden layers.The first hidden layer finds the edges, 2nd layer finds the contours, 3rd layer detects the parts of the body.The recognition process evolves from edge detection to contours detection at the next layer leading to parts detection at the next hidden layer.This in turn, leads to object detection at the last output layer.

    Figure 1:A convultional network

    The layers in CNN use the features learned by the preceding layers to recognize the larger patterns.The classification layer combines them to group the images.Its output is equal to the number of classes in the target data.The sign language used in this research is Pakistan sign language(PSL).

    The classification is done using the softmax function.The output by the softmax activation function helps in dividing each input to its corresponding classes.Accuracy is the measure of the number of true labels in the test data.Using the training data, CNN understands the object’s specific features and associates them with the corresponding category.Layers get data from the previous layer, process it, and pass it on.The network learns features of images on its own.The entire cycle starts with capturing input video, dividing it into frames of order 1280 × 720, and selecting the keyframes.The input is human action in the form of a video.The video is converted to sequential framesf1f2f3...fn.The system selects keyframes using the Median of Entropy of Mean Frames Method [43].These keyframes are recognized using CNN.

    The class to which these keyframes belong to forms an output string.The string is fed into the edit distance algorithm to find out the closest matching word.The layers in CNN use the features learned by the preceding layers to recognize the larger patterns.The classification layer combines them to group the images.Its output is equal to the number of classes in the target data.The sign language used in this research is PSL.The Input Layer is where we specify the image size for the images extracted as keyframe from the input video, which, in this case, is 234, and channel size is 1 as the images are in grayscale colors.The convolutional layer specifies filter size, which is the height and width of the filters during the training phase moved along the images extracted as a keyframe from input videos.We can use different sizes for the height and the width of the filter.Another feature is the number of filters, which specifies the number of neurons connecting to the same output area.This convolution layer determines the number of feature maps.The strides are taken as 1 for the convolution layer.The learning rate for this layer is kept relatively low.

    Fig.2 shows a complete architecture of the process of human action recognition.The ReLU Layer introduces nonlinearity in the neural network.The network uses max pooling for downsampling operation to reduce the number of parameters.This layer returns the maximum of a region of 2 × 2.The fully connected layer follows all the convolutional layers.All neurons in a fully connected layer are connected to the neurons in the previous layer.The last fully-connected layer combines them to classify the images.The fully connected layer usually uses the softmax activation function for classification.The last layer is the classification layer which uses the probabilities returned by the softmax activation function for each input to determine the output classes.The results also show the mini-batch loss and the mini-batch accuracy for the first iteration,last iteration,and every 50 iterations in between.The mini-batch loss is also called the crossentropy loss.The mini-batch accuracy is the percentage of images in the current mini-batch that the network being trained correctly classifies.It also returns the cumulative time it takes for training.Testing accuracy is a measure of the number of true labels in the test data.Using the training data,CNN understands the object’s specific features and associates them with the corresponding category.Layers get data from the previous layer,process it,and pass it on.The network learns features of images on its own,and we have no role in that.

    Figure 2:System architecture for human action recognition

    The constrainta(α1(t)+α2(t))=a(α1(t))+a(α2(t))ensures linearity and the constraint α(t)=α(t,t0)ensures time invariance.Here an input function α(t)is combined with a function η(t)to generate output that signifies convolution as a mathematical operation performed on Linear Time-Invariant (LTI) systems.The overlapping α(t) and the reverse of η(t) function.This function η(t) is the filter or kernel transformation.We define the output β as follows:

    The inputs are zero-padded at the edges,to help filters fit near the edges.The number of zeros involved in zero-padding units is another hyperparameter to improve efficiency.

    It should also be ensured to match the number of channels in the filters as well as the number of channels in its input.The convolution layer outputs go into a nonlinear layer/stage, which is just like the activation function.The detector layer normally uses the sigmoid function or hyperbolic tangent tanh ReLU for inducing nonlinearity in the model.A CNN block consists of one convolutional layer,an activation function like ReLU,and a pooling layer combined to form a network layer.The output of these blocks is flattened and sent to a fully connected output layer.

    3.1 Algorithmic Solution for Image Recognition

    Create a datastore of images StoreXi jin subfolders in datastore For each image I in a subfolder,convert all images from RGB to grayscale.Resize each image to size J× K.

    The algorithm is as under:

    Trainingpercentage ←81

    Testingandvalidationpercentage ←19

    ImageInputLayer ←1

    MaxPooling2dLayer ←1

    ClassificationLayer ←1

    Filtersize ←f

    Number of filters ←n

    Epochs ←defined n of epochs

    learningrate ←.00001

    TraintTheSystem()

    Accuracy ←trainingimagesmatched ÷ totalimages

    Misrate ←trainingimagesmismatched ÷totalimages

    ? The input layer uses only grayscale pixel values and is of sizes as shown in Tab.1.

    ? A filter of 5 with no padding and stride s=1 is used.A maxpooling layer with pool size is used with a pooling stride=2 for maxpooling.The experiment is repeated will filter of size 3×3,5×5 and 7×7.As the images are in grayscale,channel size is one.The number of filters are varied from 20 to 30 in the experiment.A fully connected layer follows it.

    ? ReLu function is used to introduce nonlinearity in the model.

    ? A maxpooling layer is introduced with stride =2.

    ? At the end of the network,a softmax layer and a classification layer is used to determine cross-entropy loss for the proposed solution.

    ? The learning rate is kept as low as 0.00001.

    3.2 Video Dataset

    Several video datasets are publicly available,however,for this research,the video dataset is formed by selecting 20 words from Pakistan sign language.15 signers participated in preparing videos for 20 words.Approximately 400 videos for every word gesture are collected.The videos are preprocessed and passed through the video summarization process.The converted images will be stored in subfolders under that word gesture folder.The images are in grayscale and of size 234 × 234.This dataset has 8,000 video clips from 20 different categories prepared for 20 words by 15 different signers.The duration of these is 11.2 hours approximately.We use these 400×20 videos by 15 signers to train,validate,and test the network.

    The video recognition process has two main components:Video summarization and image recognition.The process of video summarization consists of selecting keyframes,meaningful clip selection,and output generation.The technique proposed here uses the concept of mean and then median of entropy.The mean is a very important measure in digital image processing.It is used in spatial filtering and is helpful in noise reduction.The mean of k frames is defined as:

    Table 1:Result of applying different layers of CNN

    Here ?(i,j)shows mean of k images of size N× N.

    A video summary is generated as under:

    ? The input video is the video that is to be used for video summarization, the video may be in any standard format.

    ? Frame extraction from videos as a finite number of still images called frames.

    ? The feature extraction process can be based on features like color, edge, or motion features.Some algorithms use other low-level features such as color histogram, frame correlation, and edge histogram.

    The video is summarized to keyframes by using the technique Median of Entropy of Mean Frames Method [43].The proposed algorithm can be used for any type of video however it has performed well for continuous gestures.

    3.3 Edit Distance

    The Levenshtein distance or edit distance was named after his inventor, Vladimir I.Levenshtein.The Levenshtein distance is the number of edits needed to convert a sequence A into another sequence B.Edit operation consists of substitutions, insertions, and deletions.The variation Damerau Levenshtein distance adds an extra root in dynamic programming.It computes ad[I, j] which stores the edit distance betweena1...aiand edit operation which is a transposition.This operation interchanges two adjacent characters.To express the Damerau Levenshtein distancedbetween two strings a and b.

    4 Video Recognization

    The PSL is a very rich sign language.It consists of thousands of words.The video classification works precisely like the image classification,as explained in Section 3.1.The video is summarized,and individual frames are stored in that particular video category and frame number.At the same time, the labels of summarized frames are stored in repository DS.These frames are trained using the images in the dataset.The process is repeated for all the words in the dictionary, the summarized images are combined into folders containing similar images.The CNN model is trained by using frames in the dataset.In the test phase, the frames are used to predict the folder label.In test mode, every frame from the video summary is predicted for the category to which it belongs to.The output string consisting of the folder labels is compared with the strings in DS.The string with minimum edit distance is chosen as the output string.Algorithm of the recognized dynamic gesture given below.

    Algorithm:Recognize the Dynamic Gesture

    Input:The Video converted tof1,f2,f3...fi where1 ≤t≤tkfr

    The datastore:DS[m] dictionary of m words containing at most size number of images

    datastore:dgestures containing L folders

    Output:wordrecognized

    v[i]=label(fi)using algorithm in Section3.1

    ?fi where1 ≤i≤tk f r

    comparevwithds[m]using EditDistance?1 ≤m≤size

    wordrecognized=DS(min(ED(DS,v)))

    5 Discussion

    The dynamic sign recognition starts with image recognition.The labels of the recognized images help in identifying the dynamic gesture class, i.e., complete words using edit distance algorithm.The video recognition is analyzed as under:

    5.1 Images

    Tab.1 shows the results of applying CNN to the dataset of hand gesture images.After going through numerous training and testing rounds of the system,the accuracy of image recognition is 91.03%as shown in Tab.1 at 100 epochs,image size 234×234,and a dataset size equal to 400 which is a very reasonable rate of recognition.It can give much better results for the larger dataset and a higher number of epochs.

    Fig.3 shows a graph of image size,epochs,data set size,and accuracy.It shows that for fixed image size,the graph of the epoch,data set size and accuracy.Fig.3 shows the proposed solution accuracy with respect to dataset size&Epochs.It is shown that if epochs are increased&the dataset size is kept constant,then the accuracy of the proposed solution is increased up to 91.01%.It observed that if dataset size is increased and so are the Epochs, system accuracy is also increased.

    Figure 3:Accuracy of the proposed solution for fixed dataset size,variable epochs and variable image size

    Fig.4 shows the proposed solution accuracy with respect to constant image size (72 × 72 ), varying epoch & no of images in the data set.it observed that the accuracy of the proposed algorithm is increased with increase in epochs and no of images in the data set.Fig.4 we plotted for fixed dataset size, the graph of epochs,and image size against accuracy.

    Fig.5 shows that we get better results with higher image size.In Fig.5,the proposed method gives,for fixed Epoch=15,variable image size and variable data size,higher accuracy for higher resolution images,and higher data size with lower missrate.

    Figure 5:Accuracy of the proposed solution for fixed epoch=15,variable image size,and variable dataset size

    Every recognition system faces 4 major problems:shadow,rotation,scaling,and mirror images.Every recognition system must handle all these problems one by one.However If we train the system on images and If we can use a dataset of appropriate size, all these problems are automatically taken care of by the convolutional neural networks.

    5.2 Edit Distance

    An Edit Distance is the number of edits needed to convert,a sequence A into another sequence B.The output V from the algorithm“Recognize the Dynamic Gesture”in Section 4,is compared with all the words in the data store,and we choose the string with minimum edit distance.

    Correct Classification Rate (CCR)=×100 where,Sc,Tsrepresents the total number of samples recognized correctly using Edit Distance & the total number of samples, respectively.

    Miss Classification Rate (MCR)=×100 whereSicrepresents the total number of samples recognized incorrectly using edit distance.The following is the results from the edit distance algorithm:

    Tab.2 presents results for calculating edit distance for words of length 3, 4, 5, 6, and 7 characters respectively.

    5.3 Videos

    In PSL gestures are usually 2-5 seconds long.The videos used to form DS are converted to images after passing through the video summarization process.The image labels are stored along with the video label in the dataset DS.The video recognition process starts with the input of a gesture by the signer.CNN recognizes these images.As the images are quite complicated so learning rate is kept very low and returned labels are stored in the form of a string sequences1s2s3...snat a particular location.Tab.3 tells us about the impact of increasing epochs on mini-batch accuracy for 200 images per label.

    Table 2:Edit distance performance

    Table 3:Proposed solution performance w.r.t.epochs and minibatch accuracy & misrate

    Fig.6 shows the relationship between epochs, mini-batch accuracy, and mini-batch loss.The graph shows that with the increase of the number of epochs, the mini-batch accuracy increases and mini-batch loss decreases.At epochs = 50 the proposed system gives a mini-batch loss of 18% approximately.The same decreases to approximately 4%when epochs become equal to 200.

    The results can be improved by changing many factors,including the number of images per label,image resolution,learning rate,filter size and number of filters.As an example to test the input video,the following words were chosen:Cap,skirt,scarf,and gloves.The video title and the summarized image labels are stored in DS.This is done for all the selected videos.For this research,almost 1000 words are selected.However,more words can be added in the data store DS with an increased cost in terms of training time and a little impact on testing time.As some of the gestures are repeated, so a total number of classes, i.e., image labels, do not exceed a limit.Let’s now test an input video, V, summarized to framesf1f2f3...fn.The images are sequentially compared the edit distance algorithm isO(n2) for comparing two strings.

    Figure 6:Epochs,minibatch accuracy%, and minibatch loss%

    The computational complexity of deep neural networks is determined by matrix multiplication,nonlinear transformation and weight sharing.Dropout helps in keeping the computational complexity in the polynomial-time domain.The training part of the proposed solution is the most time-consuming part.It takes hours to get training results, however, the testing is of the order of a second.The system gives an accuracy of 90.03% on training data.The edit distance algorithm gives an accuracy of 99.99%.For the subset of words selected, it was found to be 99.74%, so the proposed system gives an accuracy of 90.79% on training data.This accuracy can be increased well above 91% by increasing the number of images per class in the dataset,increasing image resolution and increasing number of epochs.

    6 Comparison with Existing Techniques

    Table 4:Comparison of some existing techniques

    The results of the proposed solution were also compared with other existing techniques.The proposed technique achieves accuracy comparable to those provided by[25-27,29,44].Tab.4 shows the comparison of the proposed technique with other techniques.The proposed method performs reasonably well in terms of this metric.

    7 Conclusion and Future Work

    The research is an effort to facilitate the deaf society and to provide an efficient touch-free interface to users of smart devices.The proposed technique has the edge that it gives good accuracy in a constraint-free environment.The proposed methodology provides a framework for sign language recognition that can be materialized for any sign language.A larger dataset can also give better video recognition accuracy.A better algorithm for string matching of the combined output of the image recognition algorithm, which gives improved results over edit distance, is left as future work.A detailed complexity analysis of the system has also been left as future work.

    Acknowledgement:Thanks to our families & colleagues, who supported us morally.

    Funding Statement:The author(s) received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    九色亚洲精品在线播放| 三级国产精品欧美在线观看| av免费观看日本| 最黄视频免费看| 国产高清国产精品国产三级| 丝袜脚勾引网站| 老女人水多毛片| 老司机影院毛片| 欧美日韩精品成人综合77777| 亚洲精品国产色婷婷电影| 亚洲国产色片| 亚洲欧美色中文字幕在线| 看十八女毛片水多多多| 哪个播放器可以免费观看大片| 日韩强制内射视频| 国产乱来视频区| 夜夜爽夜夜爽视频| 91久久精品电影网| 一边亲一边摸免费视频| 午夜激情av网站| 亚洲欧美色中文字幕在线| 99国产综合亚洲精品| 精品久久久久久久久亚洲| 亚洲国产精品一区三区| 少妇 在线观看| 国产淫语在线视频| 永久网站在线| 亚州av有码| 亚洲国产最新在线播放| 午夜福利视频在线观看免费| 欧美xxⅹ黑人| 精品久久久久久久久av| 精品人妻熟女av久视频| 热99久久久久精品小说推荐| 国产极品天堂在线| 国产成人av激情在线播放 | 亚洲无线观看免费| 美女主播在线视频| 91精品伊人久久大香线蕉| 蜜桃国产av成人99| 亚洲精华国产精华液的使用体验| 久久久亚洲精品成人影院| 免费久久久久久久精品成人欧美视频 | 国产亚洲最大av| 性色avwww在线观看| 久久精品国产亚洲av天美| 欧美日本中文国产一区发布| 亚洲色图 男人天堂 中文字幕 | 插阴视频在线观看视频| 老熟女久久久| 狠狠婷婷综合久久久久久88av| 一级毛片 在线播放| av电影中文网址| 亚洲精品一区蜜桃| 免费观看在线日韩| 国产男人的电影天堂91| av电影中文网址| 午夜福利视频在线观看免费| 亚洲国产精品一区二区三区在线| 成人国产av品久久久| 久久精品熟女亚洲av麻豆精品| 成年美女黄网站色视频大全免费 | 热99久久久久精品小说推荐| 一区在线观看完整版| av在线播放精品| 久热久热在线精品观看| 亚洲成人av在线免费| 99热国产这里只有精品6| 免费大片18禁| 国产在线一区二区三区精| 国产精品一区二区在线观看99| 视频区图区小说| 黄色毛片三级朝国网站| 丝瓜视频免费看黄片| videosex国产| 日韩三级伦理在线观看| 高清午夜精品一区二区三区| 一本色道久久久久久精品综合| 插阴视频在线观看视频| 精品视频人人做人人爽| 欧美精品一区二区大全| 一级二级三级毛片免费看| 国产视频内射| 看免费成人av毛片| 我的女老师完整版在线观看| 最近的中文字幕免费完整| 久久婷婷青草| 超色免费av| 国产高清国产精品国产三级| 欧美日韩亚洲高清精品| 亚洲经典国产精华液单| 国产一区亚洲一区在线观看| 国产老妇伦熟女老妇高清| 男女无遮挡免费网站观看| 国产精品久久久久久久电影| 一级片'在线观看视频| 亚洲国产精品成人久久小说| h视频一区二区三区| 国产永久视频网站| 成人影院久久| 狂野欧美白嫩少妇大欣赏| 久久久欧美国产精品| 免费播放大片免费观看视频在线观看| 色哟哟·www| 51国产日韩欧美| 国产不卡av网站在线观看| 日产精品乱码卡一卡2卡三| 中国国产av一级| 亚洲av综合色区一区| 看免费成人av毛片| 中文精品一卡2卡3卡4更新| 下体分泌物呈黄色| 欧美激情国产日韩精品一区| 一本—道久久a久久精品蜜桃钙片| 亚洲四区av| 男女无遮挡免费网站观看| 亚洲精品乱码久久久久久按摩| 制服人妻中文乱码| 久久精品国产亚洲网站| 亚洲成人av在线免费| 观看av在线不卡| 精品久久久久久久久av| 国产无遮挡羞羞视频在线观看| 十八禁高潮呻吟视频| 夫妻性生交免费视频一级片| 亚洲欧洲精品一区二区精品久久久 | 日韩欧美一区视频在线观看| 日韩中字成人| 免费日韩欧美在线观看| kizo精华| 日韩不卡一区二区三区视频在线| a 毛片基地| 91精品一卡2卡3卡4卡| 亚洲色图 男人天堂 中文字幕 | 亚洲人与动物交配视频| .国产精品久久| 爱豆传媒免费全集在线观看| 国产亚洲一区二区精品| 一级毛片黄色毛片免费观看视频| 久久人人爽av亚洲精品天堂| 在线观看人妻少妇| 欧美日韩成人在线一区二区| 亚洲高清免费不卡视频| 两个人免费观看高清视频| 久久99精品国语久久久| 精品久久久精品久久久| 午夜视频国产福利| 日本午夜av视频| av国产久精品久网站免费入址| 少妇精品久久久久久久| 亚洲精品久久久久久婷婷小说| 日本av手机在线免费观看| 三上悠亚av全集在线观看| 九九在线视频观看精品| 制服人妻中文乱码| 九九爱精品视频在线观看| 亚洲欧美精品自产自拍| 天堂8中文在线网| 亚洲精品乱码久久久v下载方式| 精品国产露脸久久av麻豆| 欧美国产精品一级二级三级| 国产精品蜜桃在线观看| 亚洲精品aⅴ在线观看| av天堂久久9| 熟妇人妻不卡中文字幕| 大陆偷拍与自拍| 国产男女内射视频| 亚洲av在线观看美女高潮| 大话2 男鬼变身卡| 国产av国产精品国产| 亚洲av成人精品一区久久| 久久久久久久久久久久大奶| 亚洲成人手机| 国产不卡av网站在线观看| 又大又黄又爽视频免费| 一边摸一边做爽爽视频免费| 久久国产精品大桥未久av| 色哟哟·www| 国产欧美日韩一区二区三区在线 | 超碰97精品在线观看| 国产成人精品在线电影| 美女xxoo啪啪120秒动态图| 久久精品国产鲁丝片午夜精品| 中文天堂在线官网| 亚洲婷婷狠狠爱综合网| 另类亚洲欧美激情| 桃花免费在线播放| 香蕉精品网在线| 一级毛片黄色毛片免费观看视频| 久久99热这里只频精品6学生| 国产高清三级在线| 99热国产这里只有精品6| 日韩亚洲欧美综合| 免费av不卡在线播放| 国产男女内射视频| 亚洲图色成人| 男女边摸边吃奶| 国产高清有码在线观看视频| 热99久久久久精品小说推荐| 国产一区亚洲一区在线观看| 亚洲不卡免费看| 边亲边吃奶的免费视频| 高清不卡的av网站| 久久亚洲国产成人精品v| 亚洲伊人久久精品综合| 亚洲国产欧美日韩在线播放| 亚洲精品视频女| 久久精品久久精品一区二区三区| 久久久久国产精品人妻一区二区| 一级爰片在线观看| 久久99一区二区三区| 18禁裸乳无遮挡动漫免费视频| 一级毛片我不卡| 蜜桃久久精品国产亚洲av| 99国产精品免费福利视频| 欧美xxⅹ黑人| 欧美日韩综合久久久久久| 搡女人真爽免费视频火全软件| 日日爽夜夜爽网站| 欧美成人午夜免费资源| 99国产综合亚洲精品| 97超视频在线观看视频| 日本欧美国产在线视频| 又黄又爽又刺激的免费视频.| 国产熟女欧美一区二区| 日韩大片免费观看网站| 在线 av 中文字幕| 青春草视频在线免费观看| 老司机影院毛片| 男女高潮啪啪啪动态图| 亚洲少妇的诱惑av| 看非洲黑人一级黄片| 精品一区二区三区视频在线| 国产熟女欧美一区二区| 毛片一级片免费看久久久久| 18+在线观看网站| 日本wwww免费看| 亚洲人与动物交配视频| 在线天堂最新版资源| 校园人妻丝袜中文字幕| 亚洲国产毛片av蜜桃av| 又大又黄又爽视频免费| 妹子高潮喷水视频| 18在线观看网站| 我要看黄色一级片免费的| 国产一区二区三区综合在线观看 | 美女cb高潮喷水在线观看| 亚洲在久久综合| 国产男女内射视频| 久久久久久久亚洲中文字幕| 国产精品秋霞免费鲁丝片| 少妇 在线观看| 国产精品久久久久久久电影| 亚洲成人手机| 岛国毛片在线播放| 18禁在线播放成人免费| 久久这里有精品视频免费| 人人妻人人澡人人看| 久久精品国产亚洲网站| a级毛片在线看网站| av国产久精品久网站免费入址| 国产成人一区二区在线| 成人影院久久| 久久久精品免费免费高清| 桃花免费在线播放| 乱人伦中国视频| 欧美亚洲 丝袜 人妻 在线| 最近手机中文字幕大全| 久久久久久久久久人人人人人人| 亚洲欧美日韩卡通动漫| 一个人看视频在线观看www免费| 午夜激情久久久久久久| 老女人水多毛片| 看非洲黑人一级黄片| 麻豆乱淫一区二区| 亚洲av中文av极速乱| 日本欧美国产在线视频| 在线天堂最新版资源| 亚洲精品av麻豆狂野| 日韩在线高清观看一区二区三区| 一级毛片电影观看| 久热这里只有精品99| av福利片在线| 国产成人免费无遮挡视频| 免费看av在线观看网站| 国产男女超爽视频在线观看| 久久人人爽人人爽人人片va| 一级黄片播放器| 亚洲精品乱码久久久v下载方式| 大香蕉久久网| 亚洲精品自拍成人| xxx大片免费视频| 久久久久网色| 国产精品不卡视频一区二区| 亚洲天堂av无毛| 国产免费一区二区三区四区乱码| 在线观看一区二区三区激情| 午夜日本视频在线| 色视频在线一区二区三区| 亚洲不卡免费看| 国产免费一级a男人的天堂| 搡老乐熟女国产| 午夜精品国产一区二区电影| 亚洲欧美一区二区三区黑人 | 日韩中文字幕视频在线看片| 内地一区二区视频在线| 欧美日韩精品成人综合77777| 国产日韩欧美视频二区| 欧美精品人与动牲交sv欧美| 婷婷色综合www| 亚洲中文av在线| 丰满迷人的少妇在线观看| 青春草视频在线免费观看| 一级毛片 在线播放| 成年av动漫网址| 制服诱惑二区| 亚洲精品456在线播放app| 汤姆久久久久久久影院中文字幕| 黄色视频在线播放观看不卡| 午夜福利网站1000一区二区三区| av电影中文网址| av有码第一页| 午夜福利视频精品| 亚洲中文av在线| 亚洲精品久久午夜乱码| 国产一区二区三区综合在线观看 | 国产69精品久久久久777片| 成人毛片a级毛片在线播放| 一本久久精品| 99国产综合亚洲精品| 91精品伊人久久大香线蕉| 亚洲国产日韩一区二区| 亚洲精品久久午夜乱码| 久久久久久久久久成人| 久久久久久久精品精品| 99re6热这里在线精品视频| 久久影院123| 成人午夜精彩视频在线观看| 精品一区在线观看国产| h视频一区二区三区| 免费不卡的大黄色大毛片视频在线观看| 99久久人妻综合| 五月伊人婷婷丁香| 亚洲国产欧美日韩在线播放| 日韩伦理黄色片| 欧美 日韩 精品 国产| 国产一区二区三区av在线| 少妇熟女欧美另类| 男女啪啪激烈高潮av片| 狂野欧美激情性bbbbbb| 国产极品天堂在线| 妹子高潮喷水视频| 婷婷色综合大香蕉| 国产一级毛片在线| 国产一区二区三区av在线| 性色avwww在线观看| 乱码一卡2卡4卡精品| 久久精品人人爽人人爽视色| 极品人妻少妇av视频| 丝袜脚勾引网站| 女人久久www免费人成看片| 国产精品三级大全| 亚洲精品国产av蜜桃| 国产成人精品无人区| 久久97久久精品| 精品一品国产午夜福利视频| 一区在线观看完整版| 精品一区二区三区视频在线| 久久久久国产网址| 久久久国产一区二区| 午夜精品国产一区二区电影| 男的添女的下面高潮视频| 久久97久久精品| 久久久精品免费免费高清| 亚洲伊人久久精品综合| 欧美日韩av久久| 成人综合一区亚洲| 亚洲精品乱码久久久久久按摩| 狠狠精品人妻久久久久久综合| 成人亚洲欧美一区二区av| 国产精品一区二区在线不卡| 精品亚洲成国产av| 免费高清在线观看视频在线观看| 精品卡一卡二卡四卡免费| 日韩成人伦理影院| 国产精品国产三级国产av玫瑰| 国产精品久久久久久精品古装| 大话2 男鬼变身卡| 黄色毛片三级朝国网站| 赤兔流量卡办理| 亚洲一级一片aⅴ在线观看| 欧美亚洲日本最大视频资源| 高清黄色对白视频在线免费看| 国产av国产精品国产| kizo精华| 久久99热这里只频精品6学生| 亚洲国产日韩一区二区| 国产免费一区二区三区四区乱码| 久久人妻熟女aⅴ| 高清欧美精品videossex| 国产精品久久久久久久电影| 亚洲国产色片| 汤姆久久久久久久影院中文字幕| 久久99热6这里只有精品| 超色免费av| 亚洲av综合色区一区| 精品久久久久久久久亚洲| 久久久久久人妻| 国产一区二区在线观看av| videos熟女内射| 国产一区亚洲一区在线观看| 成人午夜精彩视频在线观看| 国产综合精华液| 国产欧美亚洲国产| 一级毛片我不卡| 成年美女黄网站色视频大全免费 | 日韩人妻高清精品专区| 久久精品国产亚洲网站| av福利片在线| 精品亚洲成a人片在线观看| 亚洲精品色激情综合| 一级毛片aaaaaa免费看小| 久久ye,这里只有精品| 久久亚洲国产成人精品v| 校园人妻丝袜中文字幕| 国产男女超爽视频在线观看| 中文乱码字字幕精品一区二区三区| 22中文网久久字幕| 亚洲av综合色区一区| 久久精品国产自在天天线| 亚洲美女搞黄在线观看| 日韩大片免费观看网站| 精品一品国产午夜福利视频| 黄片播放在线免费| 免费av不卡在线播放| 99久久精品国产国产毛片| 亚洲av欧美aⅴ国产| 免费黄色在线免费观看| 美女主播在线视频| 久久久国产欧美日韩av| 丰满乱子伦码专区| 人妻一区二区av| 亚洲欧美日韩卡通动漫| 精品国产一区二区三区久久久樱花| 春色校园在线视频观看| 午夜日本视频在线| 纯流量卡能插随身wifi吗| 久久久久久伊人网av| 成人综合一区亚洲| 9色porny在线观看| 夫妻性生交免费视频一级片| 亚洲国产精品成人久久小说| 欧美精品一区二区免费开放| 亚洲国产精品国产精品| 色哟哟·www| 高清视频免费观看一区二区| 国产成人免费无遮挡视频| 插逼视频在线观看| 少妇人妻精品综合一区二区| 亚洲精品日韩av片在线观看| 亚洲国产精品国产精品| 亚洲精品日本国产第一区| 午夜福利网站1000一区二区三区| 亚洲五月色婷婷综合| 成人黄色视频免费在线看| 亚洲精品乱码久久久v下载方式| 我的女老师完整版在线观看| 国产午夜精品一二区理论片| 天美传媒精品一区二区| 成人亚洲欧美一区二区av| 热99国产精品久久久久久7| 婷婷色麻豆天堂久久| 亚洲成人一二三区av| 国产精品秋霞免费鲁丝片| 亚洲欧美色中文字幕在线| 久久久精品94久久精品| 日韩,欧美,国产一区二区三区| 亚洲高清免费不卡视频| 边亲边吃奶的免费视频| 三级国产精品片| 国产熟女午夜一区二区三区 | 午夜久久久在线观看| 色5月婷婷丁香| 嫩草影院入口| 人妻夜夜爽99麻豆av| 天堂俺去俺来也www色官网| 日本色播在线视频| 高清在线视频一区二区三区| 精品久久国产蜜桃| 在线精品无人区一区二区三| 在线观看免费高清a一片| 男女免费视频国产| 国产精品一二三区在线看| 看非洲黑人一级黄片| 亚洲av成人精品一区久久| 在线观看国产h片| 亚洲精品美女久久av网站| 一本大道久久a久久精品| freevideosex欧美| 黑人巨大精品欧美一区二区蜜桃 | 高清欧美精品videossex| 亚洲精品亚洲一区二区| 久久久久久久大尺度免费视频| 十分钟在线观看高清视频www| 少妇被粗大的猛进出69影院 | 亚洲精品国产av成人精品| 男女边摸边吃奶| 18禁观看日本| 国产男女内射视频| 亚洲中文av在线| av女优亚洲男人天堂| 日本色播在线视频| 亚洲欧美色中文字幕在线| 精品一区在线观看国产| 国产精品一区二区在线观看99| a 毛片基地| 日韩伦理黄色片| 精品一区二区三卡| av在线播放精品| 久久99一区二区三区| 免费黄色在线免费观看| 国产欧美日韩综合在线一区二区| 国产成人精品一,二区| 在线观看美女被高潮喷水网站| 国产av码专区亚洲av| 久久精品夜色国产| 日韩制服骚丝袜av| 亚洲精华国产精华液的使用体验| 好男人视频免费观看在线| 一级,二级,三级黄色视频| 内地一区二区视频在线| 欧美另类一区| 国产白丝娇喘喷水9色精品| 欧美日韩av久久| 国产极品粉嫩免费观看在线 | 少妇人妻 视频| 成人综合一区亚洲| 高清不卡的av网站| 久久热精品热| 乱人伦中国视频| 97在线视频观看| 国产在线免费精品| 九色亚洲精品在线播放| 亚洲怡红院男人天堂| 少妇被粗大的猛进出69影院 | 大香蕉97超碰在线| 亚洲美女黄色视频免费看| 极品少妇高潮喷水抽搐| 99九九线精品视频在线观看视频| 欧美日韩一区二区视频在线观看视频在线| 久久久精品区二区三区| 国产精品人妻久久久影院| 色5月婷婷丁香| 欧美 亚洲 国产 日韩一| 亚洲精品日韩av片在线观看| 亚洲三级黄色毛片| 国产av国产精品国产| 中文字幕人妻丝袜制服| 91精品国产九色| 国产男女内射视频| 九草在线视频观看| 熟女人妻精品中文字幕| 国产一区二区在线观看av| 熟女av电影| 精品久久久精品久久久| 日本av免费视频播放| 中文字幕久久专区| 美女福利国产在线| 久久久午夜欧美精品| 亚洲少妇的诱惑av| 国产成人av激情在线播放 | 久久 成人 亚洲| 男女啪啪激烈高潮av片| 精品人妻一区二区三区麻豆| 天堂8中文在线网| 亚洲国产精品一区二区三区在线| 午夜精品国产一区二区电影| 丰满乱子伦码专区| 国产av码专区亚洲av| 亚洲精品456在线播放app| 国产精品免费大片| 免费av不卡在线播放| 少妇 在线观看| 中国三级夫妇交换| 看免费成人av毛片| 狂野欧美白嫩少妇大欣赏| 桃花免费在线播放| 一个人看视频在线观看www免费| 亚洲国产精品成人久久小说| 精品久久久久久电影网| 欧美+日韩+精品| 简卡轻食公司| 在线观看免费高清a一片| 日韩av在线免费看完整版不卡| 涩涩av久久男人的天堂| 欧美日韩视频高清一区二区三区二| 我的老师免费观看完整版| 内地一区二区视频在线| 国产日韩一区二区三区精品不卡 | 男男h啪啪无遮挡| 欧美亚洲 丝袜 人妻 在线| 日本欧美视频一区| 精品少妇内射三级| av在线观看视频网站免费| 国产色婷婷99| 美女内射精品一级片tv| 国产一区二区三区av在线| 日韩一区二区视频免费看| 插阴视频在线观看视频| 婷婷色综合大香蕉| 蜜桃在线观看..| 国产亚洲精品久久久com| 日韩中字成人| 欧美老熟妇乱子伦牲交| a级毛片黄视频|