• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Adapted Long Short-Term Memory(LSTM)for Concurrent Human Activity Recognition

    2021-12-15 08:11:54KeshavThapaZubaerMdAbdhullaAIandYangSungHyun
    Computers Materials&Continua 2021年11期

    Keshav Thapa,Zubaer Md.Abdhulla AI and Yang Sung-Hyun

    Department of Electronic Engineering,Kwangwoon University,Seoul,139-701,Korea

    Abstract:In this era,deep learning methods offer a broad spectrum of efficient and original algorithms to recognize or predict an output when given a sequence of inputs.In current trends,deep learning methods using recent long short-term memory (LSTM)algorithms try to provide superior performance,but they still have limited effectiveness when detecting sequences of complex human activity.In this work,we adapted the LSTM algorithm into a synchronous algorithm(sync-LSTM),enabling the model to take multiple parallel input sequences to produce multiple parallel synchronized output sequences.The proposed method is implemented for simultaneous human activity recognition(HAR)using heterogeneous sensor data in a smart home.HAR assists artificial intelligence in providing services to users according to their preferences.The sync-LSTM algorithm improves learning performance and sees its potential for real-world applications in complex HAR,such as concurrent activity,with higher accuracy and satisfactory computational complexity.The adapted algorithm for HAR is also applicable in the fields of ambient assistive living,healthcare,robotics,pervasive computing,and astronomy.Extensive experimental evaluation with publicly available datasets demonstrates the competitive recognition capabilities of our approach.The sync-LSTM algorithm improves learning performance and has the potential for real-life applications in complex HAR.For concurrent activity recognition,our proposed method shows an accuracy of more than 97%.

    Keywords:Concurrent HAR;deep learning;LSTM;sync-LSTM;smart home

    1 Introduction

    Activity recognition has been one of the core topics in the field of artificial intelligence (AI)research for decades.However,the various features and classifiers designed for specific activity recognition are still not satisfactory for measuring the effectiveness of detecting complex human activities.Activity recognition has started to adopt a more complex structure and inference procedures due to deep learning methods [1].These deep learning approaches have demonstrated the potential to significantly improve the state of the art in human activity recognition (HAR) [2].The complete recognition process includes sensor data,pre-processing,feature extraction,and classifiers(model),where feature extraction and classifier algorithms play a crucial role in achieving high recognition accuracy [3].Manual feature extraction or selection is often laborious and arbitrary and lacks generalizability [4].Instead of using standard produced features,we applied automatic feature extraction or selection [5,6].Recent activity recognition studies have often focused on single individuals or single activities rather than complex,concurrent activities.Our method addresses the challenge of concurrent activity detection.

    The need to structure information in order to process a large amount of data is becoming a pervasive problem and gaining high research interest.Automatic sequence classification based on convolutional neural network (CNN) [7]and recurrent neural network (RNN) [8]methods are used to structure and process these sequences with a set of high-level representations.Deep learning methods such as long short-term memory (LSTM) [9]and bi-directional LSTM (BLSTM) [10]are used in various domains and tasks,including image processing [11],language recognition [12],and sentence processing [13].

    Despite its popularity,LSTM is unable to take multiple inputs for synchronous data and cannot produce multiple outputs directly when used as a classifier.In [14],a multiple stream neural network called parallel LSTM is presented,which can process only the synchronous data stream to produce a single output.In this paper,we present an improved version of the architecture of a synchronous neural network called synchronous LSTM (sync-LSTM),which processes multiple data streams simultaneously to detect concurrent human activity.Two or more activities running simultaneously in parallel are called concurrent activities.Most importantly,we detect complex human activity using the improved deep learning platform.Multi-binary classifiers can detect parallel activity [15,16],but this fails for many activities.The hidden Markov model (HMM) [17],condition random field [18],and various other types of machine learning approaches [19]and probability inference algorithms [20]are widely used for parallel activity detection.However,they cannot handle a large number of spatio-temporal data sequences.The method proposed here is the first so far to address these issues.

    Two requirements for human activity recognition drive this work:improving recognition accuracy and developing reliable algorithms to solve complex activity recognition problems.Therefore,our method promises to address the needs of activity recognition with heterogeneous sensors,primarily by improving performance over existing approaches.The highlights of our method are as follows.

    · We present the improved LSTM known as sync-LSTM to recognize concurrent human activity.

    · The presented concept is an adapted version of LSTM that supports parallel input and parallel output.

    · The proposed approach can structure and learn spatio-temporal features directly and automatically from the raw sensor data,without the need for manual feature extraction or selection for concurrent activity recognition.

    · This framework can likely be applied to different recognition platforms with different sensor modalities in different domains.

    · The results obtained by our proposed method show that it outperforms existing methods on recognition of concurrent activity.

    Activity recognition relies on a combination of different sensors:wearable,external,or both.We chose external sensors for the user’s comfort.We evaluate and compare the performance of our proposed approach using fully annotated real-world datasets generated by Kasteren and the Center for Advanced Studies in Adaptive Systems (CASAS).The rest of this paper is organized as follows.Related work on activity recognition and LSTM is described in Section 2.Section 3 illustrates our proposed method.The experiment setup,analysis,and evaluation are provided in Section 4.Finally,Section 5 concludes the paper.

    2 Related Work

    Early research trained shallow depth classifiers on sequences collected by a single sensor.Sensor-specific constraints overly controlled this single sensor-based system,as the data from this type of sensor was inherently insufficient for complex activity recognition.Therefore,a multi-sensor heterogeneous system was designed to address these challenges.Significantly,our system can work with multiple and heterogeneous sensors.A naive Bayes classifier is used for activity recognition with the longest sequences of sensor data sequences [21].An incremental learning approach called the dynamic Bayesian network is tested to detect differences in activities by rebuilding the previously learned models [22].Early generative deep learning methods used restricted Boltzmann machines to derive task-independent feature representations [23,24].More complex models such as CNN have been successfully used for challenging HAR tasks [25].Similarly,some decent algorithms,such as multilayer perceptron [26],support vector machine [27],decision tree [28],and an updated HMM [29],are used in classifying some types of activities.

    The discriminative models have used the independence assumption,where we learn the model parameters by optimizing the conditional likelihood rather than the joint likelihood [30].Audio [31]and video [32,33]activity recognition methods are also widely used and best suited for healthcare and remote monitoring.However,audio-visual methods have privacy issues and are complex and pervasive.In recent years,many deep learning methods have been used to work with CNN [34,35],RNN [36],and LSTM.The enhanced RNN and LSTM are widely applicable in language modeling,hand gesture recognition [37],machine translation [38],sound recognition [39],video analysis [40],and image captioning [41].As mentioned earlier,they lag behind in parallel processing.Additionally,these algorithms are least common in activity recognition.However,a significant amount of research is concerned with single or regular activity recognition,and few researchers are interested in complex activity recognition such as concurrent activity recognition.Activity recognition with a CNN-LSTM [42]structure sheds light on concurrent activity detection with multimodal sensor data,but it suffers from data redundancy.

    3 Materials and Method

    The HAR process consisted of four phases:data acquisition,preprocessing,feature extraction,and training/testing,as shown in Fig.1.In this work,concurrent activity is recognized by using Sync-LSTM.Therefore,the experimental process is divided into preprocessing and training/testing,as LSTM automatically acquires the feature vectors.

    The sensor data is pre-processed by applying filters and overlapping sliding windows of 128 time steps each.Our model operates with two fully connected and two LSTM layers of 64 units each,as shown in Fig.2.We train the model and predict the real value by setting hyperparameters such as entropy,learning rate,weight decay,and optimizer.A k-fold cross-validation technique is performed to improve and validate the output before testing.During testing,a test sensor is added without affecting the learned parameters.The detection obtained during testing is then compared with the real values,and the accuracy is calculated using the F1 score determined from HAR.We present the sync-LSTM based on a standard LSTM and describe its implementation for concurrent activity recognition.The main feature of our approach is that it can accept multiple inputs and produce multiple outputs by improving the LSTM algorithm to handle parallel structure and detect concurrent activity.The input sequences are independent and mapped into homogeneous subsets.A single-mode LSTM with concatenated input sequences is theoretically not suitable for mapping the heterogeneous input representation of parallel sequences.

    Figure 1:General system workflow of human activity recognition

    Figure 2:Proposed workflow for the recognition of concurrent human activity

    3.1 Standard LSTM

    RNNs use feedback to classify current data to neurons.This unique ability of RNN helps to find patterns with long-term dependencies.However,the vanishing gradient problem still occurs.To solve this problem,LSTM was introduced.LSTM outperforms RNN in finding long-term dependencies.Fig.3a shows the internal architecture of standard LSTM.The cell in this LSTM unit of the network consists of an input gate i(t),a forgetting gate f(t),an output gate o(t),and a memory cell c(t) that stores the information and potentially updates the output over some periods.

    The LSTM also has a peephole in the inner cells to the gates in the same cell to learn the timing of the outputs.The multiplicative equation commands each cell and gate to propagate forward.

    x(t)is the input sequence.wxi,wxf,andwxoare the input weights of the sequence associated with an input gate,a forgetting gate,and an output gate,respectively.c(t)represents the memory cell at a corresponding time,and b represents the bias voltage of the corresponding gate and cell.The hidden layerh(t)computes inputx(t)and provides outputy(t)at timet,as shown in Fig.3a.σis known as the logistic sigmoid activation function,which bounds the value between [0,1]and is mathematically expressed asσ(x)=tanhis known as the hyperbolic tangent activation function.

    Figure 3:(a) Standard LSTM;(b) proposed sync-LSTM

    3.2 Adapted LSTM:Sync-LSTM

    Let a normal input sequence bex∈RS×E×I×V×L,whereSandEare the start and end times,Iis the sensor ID,Vis the sensor value,andLis the location.Sync-LSTM takes the input of sampleswhere each data pointxltis a set of individual samplesl(l=1,2,3,...N) observed by the sensors at timet(t=1,2,3,...N) and processed asx1t∈RS1×E1×I1×V1×L1,are the hidden states,whileare concurrent activities detected at the corresponding timet.H is the composite function.The insight of the proposed sync-LSTM is shown in Fig.3b,which contains the input gateilt,the forget gateflt,the output gateolt,and the cell memoryclt.W(t) is the weight matrix.Each gate has its activation functions denoted by sigmoid (σ) and hyperbolic tangent (∫).Although synchronization is a more challenging prototype that may affect the convergence system,this paradigm can accelerate the training model and provide a viable solution to facilitate the development and operation of AI applications.

    Figure 4:Unfolded architecture of sync-LSTM

    Sync-LSTM forwards multiple activity sequences to detect and encode hidden internal structures between parallel hidden activity sequences.It is time-consuming to process a signal with long time steps by a standard LSTM neuron.Therefore,we used multiple LSTM units in parallel to process different parts of the information.Fig.4 shows the unfolded architecture of the sync-LSTM network.It consists of the input layer,the parallel LSTM,the fully connected layers,and the outputs.The outputs in the final time step of each LSTM unit are summarized as n×h,where h is the number of hidden neurons of each LSTM unit.The LSTM layers adjust their internal state after each time step.The size of the weight,bias,cell,and hidden layers is assigned to be 128.The final number of parameters depends on the number of activity classes in the classification.

    Algorithm 1:Pseudo-code for concurrent activity recognition using sync-LSTM 1.initialize network 2.reset:inputs=0,activations=0 forward propagation:3.initialize the inputs do 4.roll over:activations;cell states 5.loop over a cell,end for 6.do for t=0 to n do Calculate the gate values:input gates:ilt=σimages/BZ_233_649_1098_674_1144.pngwxi*xlt+whi*hlt-1+wci * clt-1+biimages/BZ_233_1369_1098_1393_1144.pngforget gates:flt=σimages/BZ_233_680_1181_705_1226.pngwxf *xlt+whf * hlt-1+wcf * clt-1+bfimages/BZ_233_1441_1181_1466_1226.pngloop over the cells in block now output gates:olt=σ(wxo * xlt+who * hlt-1+wco * clt-1+bo)update the cell:clt=flt * clt-1+ilttanh* (wxcxlt+whchlt-1+bc)final hidden state/final output:hlt=olt * tanh * clt hlt=H(wxlhlxlt+whlhlhlt-1+blh)Ylt=(wylhlhlt-1+bly)end for 7.concurrent activity recognize 8.do Update the weight end

    4 Experimental Results and Analysis

    This section presents the detailed results in both the training and recognition phases of the model.Several design parameters,such as the input data,the number of sync-LSTM layers,and the number of activities,are assigned and processed.The training dataset is used to train the classification,estimation,and evaluation of an individual activity for the best model parameters and for tuning the hyperparameters.Then the proposed model is trained,and the results are compared with the existing model outputs.For the experimental analysis of the proposed approach,the Kasteren and Kyoto 3 datasets are considered.The selected datasets have some limitations,such as the activity instances of different groups,some residents in each house,performance of the same activity in different ways,and availability of less learning data.Algorithm 1 presents the pseudo-code for concurrent activity recognition.

    4.1 Experimental Configuration and Training

    The proposed neural network is implemented in the TensorFlow and scikit-learn libraries.In this paper,linear interpolation is used to fill the missing data and normalized to a zero mean with a standard deviation of 0.5.The sensor data is pre-processed and sampled in overlapping sliding windows with a fixed width of 200 ms and a window size ranging from 0.25 s to 7 s.The data is sampled in a single window.The proposed method is trained and tested using the TensorFlow_GPU1.13.1 library.The computer deployment is best suited to run our algorithm using i7 CPU with 16 GB RAM and GTX Titan GPU with CUDA 9.0 and using the cuDDN 7.0 library.The CPU and GPU are used to avoid exceeding the memory limit during training.The dataset is split into a training set and a test set.The model uses 70% of the data for training,10%for validation,and the remaining 20% for testing.We use a k-fold CV (cross-validation) to validate the data.In our experiment,we validate with k=10,known as 10-fold cross-validation.The accuracy result is averaged across all 10 folds,and the error is calculated as the cross-validation error rate.

    During training,the dropout rate is set to 0.5 to remove and avoid unused specific neurons from each hidden layer to solve the overfitting problem.The training loss function is minimized by random initialization and optimization of training parameters.The two-loss functions,named cross-entropy and L2 normalization,are inherited to avoid overfitting throughout the epoch to make the model stable.

    where m is the number of samples per batch,and Γ is the weighting parameter.yltis the predicted output,andis the label from the dataset.L2-normalization limits the trainable weighting parameter to a smaller value,which avoids overfitting.

    We try to tune the best hyperparameters as shown in Tab.1 in networks so that the learning rate and L2 weight decrease to reduce the difference and thus achieve the possible optimal performance.We train the model with a learning rate of 0.005 and 0.006 for the Kyoto 3 and Kasteren House-B datasets by taking the batch value of 100 for each epoch.Learning starts at 0.001 for all data.The training is performed for more than 12,000 epochs and stops at stable outputs.The Adam optimizer is an adaptive moment estimator that obtains independent adaptive learning rates for different parameters,making them more stable.The dimension based on the input is set to 128.The gradient clipping is set to 4 and 5 to reduce the threshold for crossing the gradient to match the normalization.The batch size is set to 100 samples,and this is often referred to as a mini-batch gradient descent since our batch size is smaller than the training sample size.

    4.2 Datasets

    Our proposed method is evaluated using the Kasteren datasets [43]and CASAS [44].An overview of the datasets is shown in Tab.2.Both datasets were gathered in an apartment containing either four or two rooms.Seventy-six sensors are deployed in Kyoto 3,whereas 23 sensors are deployed in House B.Of these sensors,51 were used for motion detection,12 were used as cabinet sensors,5 were used for cooking element detection,3 were used as temperature sensors,and the remainder was used as a medicine container sensor,pot sensor,phone book sensor,water sensor,burner sensor,or phone sensor in Kyoto 3.The four residents in the Kyoto 3 dataset performed eight different activities for 15 days.There are 178 instances of activity recorded from Kyoto 3.The single resident in the Kasteren house performed 13 different types of activities for 14 days.The sensors show the change of state according to the action of the occupant.Radiofrequency identification,a wireless sensor network,a pressure sensor,a reed switch,mercury contacts,a passive infrared-PIR,float sensors,and temperature sensors are used to record the data for the Kasteren house.The schematic diagram of the sensor deployment is shown in Fig.5.

    Table 1:Hyperparameter settings

    Table 2:Overview of Kyoto 3 and Kasteren House-B dataset

    4.3 Evaluation Metrics

    The metrics of the confusion matrix,accuracy,F1-score,and training time are used to evaluate the performance of the model.A confusion matrix shows the performance of the approach,where the row represents the predicted class,and the column represents the actual class and vice versa.It also gives information about the errors made.Generally,human activity recognition methods are evaluated according to their computational recognition accuracy.The accuracy is calculated using the confusion matrix,which is the result of the Precision and Recall parameters.Precision is the proportion of correctly recognized instances out of the absolute perceived activity occurrences.Recall is the proportion of correctly recognized instances out of the total occurrences of the activity.An f-score is the weighted average of Precision and Recall whose value is between[0,1],with the best performance indicated if it is closer to 1.

    Figure 5:Schematic diagram of sensor deployed and layout (a) Kyoto 3 (b) Kasteren House-B

    where tp,tn,fp,and fn are true-positive,true-negative,false-positive,and false-negative,respectively.The true-positive score is defined as the number of true activities detected in positive instances,while a false-positive indicates the false activities detected in negative instances.The false-negative score indicates the exact number of false activities detected in positive instances,whereas the true-negative score reflects the correct non-detection of activities in the negative instances.

    4.4 Recognition Analysis

    In this section,a possible implementation of the platform for human activity detection in smart homes is explained.All activities are localized based on the dataset.The activities that occur most frequently at the same time are considered the predominant activities in a smart home environment.Fig.6 presents the graphical confusion matrix for Kyoto 3.

    Figure 6:Confusion matrix for Kyoto 3

    According to the confusion matrix,thefilling medicationactivity was correctly detected with 98% accuracy but still has an activity error of 2% because the call may be made while other activities are being performed simultaneously.Watching DVDalso has a 98% confusion accuracy,although 1.3%,1.2%,and 0.4% of thefill medication dispenser,answer the phone,andchoose outfitactivities cause confusion since they can all be performed at the same location.The activitieswater plants,answer the phone,prepare birthday card,prepare soup,clean,andchoose outfithave recognition accuracies of 98%,96%,97%,99%,98%,and 98%,respectively.The activityanswer the phonehas the lowest recognition accuracy compared to the rest of the activities,and this activity also leads to confusion with all other activities because a phone call can be performed with all other activities that share most sensor values.However,the co-occurrence ofanswer the phoneis higher.The data from Kasteren and CASAS have the lowest number of records and instances;therefore,the actual distribution is easy to find and train.The recognition accuracy can increase if we train our proposed method with a large amount of data.

    Figure 7:Confusion matrix for House-B

    The confusion matrix for House-B is shown in Fig.7.The number of activity instances is relatively small,so the occurrence of errors is relatively lower,and recognizing the concurrent activity with the highest true positive value results in 96.90% accuracy.The activitydrinkingin House-B is the most appropriate activity that occurs concurrently with the activitiespreparing dinner,preparing breakfast,dinner,relaxing,using dishwasher,andleaving house,with an accuracy of more than 96.90%.House-B achieved precision,recall,and an f1-score of 97.94%,97.00%,and 0.97,respectively.Brushing teethis also detected simultaneously with the activitytoileting.Activities such asbrushing teeth,showering,andtoiletingcreate confusion with some error because all the activities share the same location.However,the errors are comparatively very low and can be neglected.Theusing dishwasheractivity shows concurrency with activities such aspreparing dinner,preparing breakfast,dinner,andbreakfast.The detection ofusing dishwasherconcurrently with other activities is more than 97.935%.All of these concurrently recognized activities have high detection accuracies.Sleepingis a stand-alone activity:It cannot appear simultaneously with other listed activities.Sometimes it may create some confusion with thedressingactivity,as it is performed in the bedroom.The accuracy of House-B is insufficient to fully establish the experimental concept.Although the accuracy is high,many datasets could be needed to find the proposed actual recognition.We confirm that our proposed approach for concurrent human activity recognition is feasible.

    4.5 Performance Comparison

    Figs.8a and 8b present the accuracy and loss of training and testing procedures for Kyoto 3 and House-B,respectively.In the graphs,the gap between the training and testing accuracy is comparatively small,indicating the effectiveness of the model.The gap between training and test loss is also very narrow,which explains that the dropout techniques are beneficial and resistant to overfitting.

    Figure 8:Accuracy and loss curve of (a) Kyoto 3 and (b) House-B

    The average accuracy was found to be 97.374%,and the average error was 0.1637.The performance of the proposed approach was compared with the existing framework,which uses CNN,LSTM,and Bi-LSTM methods (algorithms) for recognition by measuring the average precision,recall,f-score,and accuracy,which are shown in Tab.4.The accuracy of our method is more than 97%,and the f1-score is more than 0.97.The Bi-LSTM also has competitive accuracy with our method,but it can only process two inputs simultaneously.The cross-validation process is performed before the test to validate the input sequences;therefore,the accuracy of the test increases.This is a natural phenomenon in AI.The mean and standard deviation of the accuracy and error using 10-fold cross-validation are shown in Tab.3.The given analysis shows that the proposed method can detect concurrent activities with higher accuracy than the existing approaches.

    Table 3:Average accuracy and standard deviation (SD) over the 10-fold CV

    Table 4:Accuracy comparison with existing approaches

    4.6 Computational Complexity

    The computational complexity depends on the number of weights and is given asO(W),where W is the weight.The weight depends on the number of output units,the cell storage unit,the size of the memory,and the number of hidden units.It is also affected by the number of units associated with forwarding neurons,memory cells,gate units,and hidden units.The computational complexity does not depend on the length of the input sequence.Although using an LSTM framework increases time complexity,our approach has an acceptable computation time.Unlike concurrent activity detection,Bi-LSTM has lower complexity,but it can only handle two parallel activity detection processes and apply delay or other chaining functions to recognize more than two activities.

    This feature causes the system to wait for a complete process,which increases the computational complexity.Fig.9 shows the computational time for testing our method compared to existing frameworks,such as CNN,LSTM,and Bi-LSTM,for Kasteren house and Kyoto 3.In the CNN comparison,it processes too many hidden layers and pooling as it segments the activity into other sub-activities,which causes higher computational complexity.The computation time is slightly higher but satisfactory and executable.

    Figure 9:Computation time comparison with existing approaches

    5 Conclusion

    The framework presented in this paper shows that sync-LSTM can lead to a feasible solution for detecting concurrent human activities in the smart home scenario.This claim is supported by a comprehensive comparison with recently utilized activity recognition techniques,such as CNN,LSTM,and bi-LSTM.LSTM can work with single data sequences and bi-LSTM with a maximum of two data sequences,but our sync-LSTM can accept multiple inputs and generate multiple outputs,synchronized and in parallel.That is,for any environment or domain for concurrent processing and recognition,this sync-LSTM would be an effective solution.Many approaches focus on single and regular activity detection.Few of them have tried to detect complex activity.Starting from the standard LSTM formulation,we have improvised a more efficient LSTM-based approach to recognize complex human activity.

    However,accuracy,processing complexity,and complex activity recognition are still significant challenges in human activity recognition.The proposed method has an f1 score of more than 0.97,along with an accuracy of more than 97%.This proves its effectiveness for concurrent human activity detection with successful training and testing.Nevertheless,the accuracy is limited due to some error factors,such as same location errors,sensor distance,noise interference,and limited data.The unique best-performing model also suffers from several real-time challenges across different datasets.The parameters,like number of activities,type of sensors,sensor distribution,number of occupants,and duration of test periods,also affect the performance.The window size plays a significant role in performance,as a small size may not contain all of the information,and large size may lead to resident detection errors.The proposed method processes parallel data,which is beneficial and consistent with the setting of highly imbalanced datasets.Therefore,data augmentation techniques are not required.

    Besides,sync-LSTM can automatically extract spatio-temporal information by reducing the time-consuming effort for pre-processing data and manual feature extraction.External sensors were used instead of wearables and camera or video sensors to avoid the unnecessary burden and protect the privacy of the resident.In the future,more complex activities,such as interleaved activity,will be recognized by improving and updating the proposed method.Furthermore,we can take advantage of cloud computing by using Google Colab and Amazon Web Services.These technologies provide the opportunity to use their servers and also experiment with tensor processing units.Using these techniques and technologies will also reduce the time complexity for faster and better performance.We will also explore a transfer learning approach for this model in other domains,environments,and sectors on big data and cloud infrastructures.In summary,our proposed method (i.e.,a sync-LSTM-based model that provides fewer parallel and synchronized recognition and prediction paradigms) is preferable to its competitors.

    Funding Statement:This research was supported by the Ministry of Trade,Industry &Energy of the Republic of Korea as an AI Home Platform Development Project (20009496) and conducted under a research grant from Kwangwoon University in 2021.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    日本与韩国留学比较| av国产久精品久网站免费入址| 噜噜噜噜噜久久久久久91| 欧美少妇被猛烈插入视频| 视频区图区小说| 婷婷色综合大香蕉| 免费观看a级毛片全部| 五月伊人婷婷丁香| 国产精品秋霞免费鲁丝片| 最新中文字幕久久久久| 美女国产视频在线观看| 一区二区av电影网| 爱豆传媒免费全集在线观看| 亚洲av电影在线观看一区二区三区| 欧美日韩视频高清一区二区三区二| 丰满少妇做爰视频| 国产精品99久久99久久久不卡 | 下体分泌物呈黄色| 欧美成人一区二区免费高清观看| 97热精品久久久久久| 女性被躁到高潮视频| 国模一区二区三区四区视频| 亚洲国产精品一区三区| 日韩制服骚丝袜av| av免费观看日本| 国产精品成人在线| 国产午夜精品久久久久久一区二区三区| 日韩中字成人| 精品人妻视频免费看| 自拍偷自拍亚洲精品老妇| 男女边摸边吃奶| 国产精品成人在线| 在线观看一区二区三区| 亚洲欧美成人综合另类久久久| 男女下面进入的视频免费午夜| 九九久久精品国产亚洲av麻豆| 菩萨蛮人人尽说江南好唐韦庄| 婷婷色综合大香蕉| 亚洲综合色惰| 欧美少妇被猛烈插入视频| 精品少妇久久久久久888优播| 精品人妻视频免费看| 如何舔出高潮| 国产 精品1| 在线观看免费高清a一片| 色视频在线一区二区三区| 黄片无遮挡物在线观看| 日韩av在线免费看完整版不卡| 只有这里有精品99| 亚洲熟女精品中文字幕| 亚洲熟女精品中文字幕| 色婷婷av一区二区三区视频| 26uuu在线亚洲综合色| 搡女人真爽免费视频火全软件| 少妇人妻久久综合中文| 国产av码专区亚洲av| 亚洲无线观看免费| av在线蜜桃| 日本黄大片高清| 寂寞人妻少妇视频99o| 免费观看a级毛片全部| 五月玫瑰六月丁香| 伦理电影免费视频| 亚洲av免费高清在线观看| 特大巨黑吊av在线直播| 一级爰片在线观看| 99re6热这里在线精品视频| 亚洲第一区二区三区不卡| 大话2 男鬼变身卡| 亚洲欧洲国产日韩| 久久影院123| 亚洲欧美日韩卡通动漫| 色视频在线一区二区三区| 少妇猛男粗大的猛烈进出视频| 18禁裸乳无遮挡免费网站照片| 男人和女人高潮做爰伦理| 午夜激情福利司机影院| 一级黄片播放器| 亚洲av不卡在线观看| av在线老鸭窝| 国产精品国产av在线观看| 国产一区二区三区综合在线观看 | 熟女电影av网| 美女脱内裤让男人舔精品视频| 高清日韩中文字幕在线| 丝袜脚勾引网站| 男女无遮挡免费网站观看| 人妻夜夜爽99麻豆av| xxx大片免费视频| av不卡在线播放| 又黄又爽又刺激的免费视频.| 日本-黄色视频高清免费观看| 成人国产av品久久久| 最黄视频免费看| 日本av手机在线免费观看| 欧美 日韩 精品 国产| 不卡视频在线观看欧美| 少妇被粗大猛烈的视频| 国产精品久久久久成人av| 国产精品麻豆人妻色哟哟久久| 制服丝袜香蕉在线| 亚洲欧洲日产国产| 成人二区视频| 免费大片黄手机在线观看| 免费观看a级毛片全部| av在线播放精品| 亚洲精品国产av蜜桃| 少妇高潮的动态图| 亚州av有码| 久久久色成人| 熟女av电影| 久久99蜜桃精品久久| 超碰av人人做人人爽久久| 亚洲精品第二区| 蜜桃在线观看..| 欧美xxⅹ黑人| 精品久久久久久电影网| a 毛片基地| 18禁动态无遮挡网站| 国产精品不卡视频一区二区| 精品国产三级普通话版| 国产精品嫩草影院av在线观看| 黄色欧美视频在线观看| 免费人妻精品一区二区三区视频| 久久 成人 亚洲| 免费黄网站久久成人精品| av福利片在线观看| 夫妻午夜视频| 18禁动态无遮挡网站| 人人妻人人澡人人爽人人夜夜| 菩萨蛮人人尽说江南好唐韦庄| 18禁裸乳无遮挡免费网站照片| 欧美一级a爱片免费观看看| 永久免费av网站大全| 精品久久国产蜜桃| 婷婷色av中文字幕| 夫妻性生交免费视频一级片| 有码 亚洲区| 天堂俺去俺来也www色官网| 精品一品国产午夜福利视频| 国产淫语在线视频| 免费观看性生交大片5| 一级毛片久久久久久久久女| 免费大片18禁| 久久人人爽人人爽人人片va| 少妇被粗大猛烈的视频| 91久久精品国产一区二区三区| 一区二区三区乱码不卡18| 中文字幕制服av| 日本欧美视频一区| av天堂中文字幕网| av免费观看日本| 日日摸夜夜添夜夜爱| 亚洲精品成人av观看孕妇| 熟女av电影| 欧美日韩视频精品一区| 成人18禁高潮啪啪吃奶动态图 | 精品久久久精品久久久| 97超视频在线观看视频| 自拍欧美九色日韩亚洲蝌蚪91| 人人妻人人添人人爽欧美一区卜| 9191精品国产免费久久| 国产视频一区二区在线看| 亚洲欧美一区二区三区国产| 亚洲精品日本国产第一区| 精品高清国产在线一区| 国产成人av激情在线播放| 精品一区在线观看国产| 国产男女内射视频| 亚洲精品乱久久久久久| 亚洲国产av影院在线观看| 制服诱惑二区| 亚洲色图 男人天堂 中文字幕| 亚洲人成电影观看| 亚洲欧洲日产国产| 亚洲一区中文字幕在线| 免费少妇av软件| 美女高潮到喷水免费观看| 亚洲精品自拍成人| www.精华液| 精品国产国语对白av| 亚洲精品第二区| 国产成人系列免费观看| 欧美xxⅹ黑人| 色婷婷av一区二区三区视频| 黑人欧美特级aaaaaa片| 国产精品一二三区在线看| 午夜影院在线不卡| 中文字幕另类日韩欧美亚洲嫩草| 黄色片一级片一级黄色片| 男女边吃奶边做爰视频| 婷婷色av中文字幕| 人成视频在线观看免费观看| 国产成人免费无遮挡视频| 国产又爽黄色视频| 观看av在线不卡| 一区二区三区乱码不卡18| 考比视频在线观看| 七月丁香在线播放| 蜜桃国产av成人99| 人成视频在线观看免费观看| 亚洲成人手机| 一区二区日韩欧美中文字幕| 国产麻豆69| 午夜激情av网站| 久久鲁丝午夜福利片| 亚洲免费av在线视频| 亚洲欧美中文字幕日韩二区| 色94色欧美一区二区| 欧美日韩亚洲国产一区二区在线观看 | 搡老乐熟女国产| 国产欧美日韩一区二区三区在线| 视频在线观看一区二区三区| 丝袜美足系列| 亚洲国产精品999| 亚洲av成人精品一二三区| 国产精品熟女久久久久浪| 欧美黑人欧美精品刺激| 啦啦啦在线观看免费高清www| 免费看av在线观看网站| 大型av网站在线播放| 青青草视频在线视频观看| 国产亚洲av片在线观看秒播厂| 日韩人妻精品一区2区三区| 亚洲,一卡二卡三卡| 欧美精品高潮呻吟av久久| 免费高清在线观看视频在线观看| 热99国产精品久久久久久7| 欧美人与善性xxx| 色婷婷av一区二区三区视频| 免费在线观看影片大全网站 | 国产免费视频播放在线视频| 亚洲人成网站在线观看播放| 国产97色在线日韩免费| 婷婷成人精品国产| 一级,二级,三级黄色视频| 男男h啪啪无遮挡| 精品视频人人做人人爽| 七月丁香在线播放| 欧美日韩亚洲高清精品| 免费看十八禁软件| 超碰成人久久| 日韩熟女老妇一区二区性免费视频| 国产男女内射视频| 在线观看人妻少妇| av福利片在线| 国产亚洲一区二区精品| 交换朋友夫妻互换小说| 婷婷色综合www| 人妻一区二区av| 一本色道久久久久久精品综合| 亚洲熟女毛片儿| 国产亚洲精品久久久久5区| 国产精品一区二区免费欧美 | 性色av乱码一区二区三区2| 天堂8中文在线网| av在线app专区| 超色免费av| 一本大道久久a久久精品| 黄片小视频在线播放| 国产成人欧美在线观看 | 国产国语露脸激情在线看| 手机成人av网站| 大香蕉久久网| 欧美久久黑人一区二区| 少妇粗大呻吟视频| 国产精品偷伦视频观看了| 亚洲中文日韩欧美视频| 日本黄色日本黄色录像| 久久国产精品人妻蜜桃| 午夜福利一区二区在线看| 午夜老司机福利片| 欧美成人精品欧美一级黄| 久久免费观看电影| 老司机深夜福利视频在线观看 | 99精品久久久久人妻精品| 手机成人av网站| 午夜福利在线免费观看网站| 一级毛片女人18水好多 | 久久精品久久精品一区二区三区| 亚洲精品一二三| 成人国语在线视频| 老汉色∧v一级毛片| 男的添女的下面高潮视频| 99九九在线精品视频| 欧美黑人欧美精品刺激| av天堂久久9| 一个人免费看片子| 日韩一卡2卡3卡4卡2021年| 久久久久久免费高清国产稀缺| 欧美xxⅹ黑人| 中文字幕av电影在线播放| 久久亚洲精品不卡| 亚洲精品乱久久久久久| 韩国精品一区二区三区| 在线看a的网站| 啦啦啦在线观看免费高清www| 欧美在线一区亚洲| 9色porny在线观看| 嫩草影视91久久| 叶爱在线成人免费视频播放| svipshipincom国产片| 精品少妇一区二区三区视频日本电影| www.av在线官网国产| 老司机午夜十八禁免费视频| 九色亚洲精品在线播放| 欧美激情极品国产一区二区三区| 国产精品人妻久久久影院| 下体分泌物呈黄色| 国产精品九九99| 女人精品久久久久毛片| 一级黄片播放器| 欧美日本中文国产一区发布| 爱豆传媒免费全集在线观看| 国产精品国产三级国产专区5o| 99国产精品免费福利视频| 欧美日韩福利视频一区二区| 亚洲av综合色区一区| 国产黄色免费在线视频| 精品少妇一区二区三区视频日本电影| 日韩av在线免费看完整版不卡| 国产精品熟女久久久久浪| 色网站视频免费| av国产精品久久久久影院| 亚洲视频免费观看视频| av天堂久久9| 午夜免费观看性视频| 亚洲成国产人片在线观看| 国产在线一区二区三区精| 欧美日韩视频精品一区| 国产一级毛片在线| 日本午夜av视频| 另类精品久久| 日本色播在线视频| 婷婷色av中文字幕| 日本色播在线视频| 亚洲国产成人一精品久久久| 别揉我奶头~嗯~啊~动态视频 | 亚洲久久久国产精品| 久久ye,这里只有精品| 男人添女人高潮全过程视频| 亚洲国产欧美在线一区| 国产日韩欧美在线精品| 啦啦啦在线观看免费高清www| 免费黄频网站在线观看国产| 国产精品亚洲av一区麻豆| 亚洲av成人不卡在线观看播放网 | 亚洲色图综合在线观看| 岛国毛片在线播放| 免费一级毛片在线播放高清视频 | av欧美777| 一本色道久久久久久精品综合| 亚洲国产精品国产精品| 一级毛片 在线播放| 国产一级毛片在线| 在线观看免费午夜福利视频| 国产欧美日韩一区二区三区在线| 视频区图区小说| 国产欧美日韩一区二区三区在线| 人人妻人人澡人人看| 中文字幕制服av| 国产熟女午夜一区二区三区| 中文字幕制服av| 69精品国产乱码久久久| 女人被躁到高潮嗷嗷叫费观| 视频区图区小说| 999久久久国产精品视频| 日本色播在线视频| 国产爽快片一区二区三区| 日本午夜av视频| 免费在线观看影片大全网站 | 亚洲精品第二区| 一级,二级,三级黄色视频| 九色亚洲精品在线播放| 亚洲人成电影观看| 国产精品久久久久久精品古装| 久久免费观看电影| 亚洲三区欧美一区| 国产熟女午夜一区二区三区| 自线自在国产av| 久久这里只有精品19| 成人国产一区最新在线观看 | 美女中出高潮动态图| 最新在线观看一区二区三区 | 亚洲欧洲日产国产| 久久久欧美国产精品| 国产精品欧美亚洲77777| 久久久精品免费免费高清| 免费av中文字幕在线| 午夜福利乱码中文字幕| 激情五月婷婷亚洲| 高清视频免费观看一区二区| 日日爽夜夜爽网站| 人人妻人人澡人人爽人人夜夜| 99久久综合免费| 夫妻性生交免费视频一级片| 美国免费a级毛片| 欧美激情高清一区二区三区| 2021少妇久久久久久久久久久| 亚洲七黄色美女视频| 久久久久网色| 国产黄色免费在线视频| 蜜桃国产av成人99| 国产成人av教育| av天堂在线播放| 女人精品久久久久毛片| 久久狼人影院| 国产精品一区二区在线不卡| 国产精品二区激情视频| 高清不卡的av网站| 最近中文字幕2019免费版| 中文字幕精品免费在线观看视频| 国产精品久久久久久人妻精品电影 | 日韩一卡2卡3卡4卡2021年| 新久久久久国产一级毛片| 狂野欧美激情性xxxx| 久久国产精品男人的天堂亚洲| 一边亲一边摸免费视频| 热99久久久久精品小说推荐| 精品少妇久久久久久888优播| 黄色一级大片看看| 亚洲欧美一区二区三区久久| 精品免费久久久久久久清纯 | 国产黄频视频在线观看| 悠悠久久av| 美女主播在线视频| 国产在线一区二区三区精| 久久久久久亚洲精品国产蜜桃av| 久久精品熟女亚洲av麻豆精品| 一级毛片电影观看| 人妻一区二区av| 亚洲,一卡二卡三卡| 90打野战视频偷拍视频| 欧美国产精品va在线观看不卡| 男女下面插进去视频免费观看| 国产精品免费视频内射| 麻豆国产av国片精品| 男人舔女人的私密视频| 视频区图区小说| 国产无遮挡羞羞视频在线观看| 男的添女的下面高潮视频| 少妇猛男粗大的猛烈进出视频| 一级毛片 在线播放| 一级黄色大片毛片| 大型av网站在线播放| av国产久精品久网站免费入址| videosex国产| 黄色a级毛片大全视频| 亚洲精品一二三| 青春草亚洲视频在线观看| 高清av免费在线| 国产成人精品久久二区二区91| 在线观看免费高清a一片| 国产在线免费精品| 久久国产精品人妻蜜桃| 久热这里只有精品99| 七月丁香在线播放| 国产一级毛片在线| 这个男人来自地球电影免费观看| www日本在线高清视频| 视频区欧美日本亚洲| 国产99久久九九免费精品| 欧美变态另类bdsm刘玥| 91麻豆av在线| 国产日韩欧美亚洲二区| 别揉我奶头~嗯~啊~动态视频 | 精品卡一卡二卡四卡免费| 久久精品国产亚洲av涩爱| 成人影院久久| 精品视频人人做人人爽| 狂野欧美激情性xxxx| 十八禁高潮呻吟视频| 69精品国产乱码久久久| 啦啦啦啦在线视频资源| 久久精品成人免费网站| 曰老女人黄片| cao死你这个sao货| 丰满饥渴人妻一区二区三| 亚洲欧美成人综合另类久久久| 国产精品久久久av美女十八| 国产视频首页在线观看| 黄色视频不卡| 免费日韩欧美在线观看| 久久精品亚洲av国产电影网| 女人高潮潮喷娇喘18禁视频| 亚洲国产中文字幕在线视频| 日韩,欧美,国产一区二区三区| 精品国产国语对白av| 欧美精品一区二区大全| 欧美 日韩 精品 国产| 在线av久久热| 亚洲人成网站在线观看播放| 久久国产亚洲av麻豆专区| 大话2 男鬼变身卡| 亚洲av综合色区一区| e午夜精品久久久久久久| h视频一区二区三区| 91老司机精品| 99精国产麻豆久久婷婷| 亚洲欧美日韩另类电影网站| 99热国产这里只有精品6| 欧美精品高潮呻吟av久久| 久久午夜综合久久蜜桃| 看十八女毛片水多多多| 国产片特级美女逼逼视频| 日本av免费视频播放| 精品亚洲成a人片在线观看| av一本久久久久| 国产在线免费精品| 在线 av 中文字幕| 午夜av观看不卡| 91成人精品电影| 男女国产视频网站| 19禁男女啪啪无遮挡网站| 高清欧美精品videossex| 久久精品成人免费网站| 自线自在国产av| 高清av免费在线| 精品少妇黑人巨大在线播放| 亚洲欧美清纯卡通| 又大又爽又粗| 久久人人97超碰香蕉20202| 成人三级做爰电影| 十八禁网站网址无遮挡| 欧美 亚洲 国产 日韩一| 麻豆av在线久日| 天天躁夜夜躁狠狠躁躁| 日韩av免费高清视频| 亚洲精品美女久久久久99蜜臀 | 国产精品久久久久久人妻精品电影 | 国产人伦9x9x在线观看| 免费高清在线观看日韩| 秋霞在线观看毛片| e午夜精品久久久久久久| 1024香蕉在线观看| 欧美日韩国产mv在线观看视频| 免费看十八禁软件| 97在线人人人人妻| 久久国产精品人妻蜜桃| 欧美激情高清一区二区三区| 妹子高潮喷水视频| 久久久久国产一级毛片高清牌| 亚洲久久久国产精品| 欧美 亚洲 国产 日韩一| 制服诱惑二区| 美女午夜性视频免费| 人人妻人人添人人爽欧美一区卜| 亚洲欧美一区二区三区黑人| 国产无遮挡羞羞视频在线观看| 爱豆传媒免费全集在线观看| 在线精品无人区一区二区三| 五月天丁香电影| 99香蕉大伊视频| 国产xxxxx性猛交| 亚洲,一卡二卡三卡| 亚洲av欧美aⅴ国产| 亚洲 国产 在线| 亚洲国产中文字幕在线视频| 90打野战视频偷拍视频| 国产日韩欧美视频二区| 日韩大片免费观看网站| 欧美变态另类bdsm刘玥| 91九色精品人成在线观看| 久久精品亚洲av国产电影网| 国产精品久久久av美女十八| 久久精品久久精品一区二区三区| 精品福利永久在线观看| 国产爽快片一区二区三区| www.熟女人妻精品国产| 国产精品av久久久久免费| 午夜两性在线视频| 午夜老司机福利片| 欧美老熟妇乱子伦牲交| 99国产精品一区二区三区| 别揉我奶头~嗯~啊~动态视频 | 免费看十八禁软件| 极品少妇高潮喷水抽搐| 精品第一国产精品| 大话2 男鬼变身卡| 亚洲 国产 在线| av天堂久久9| 性色av乱码一区二区三区2| 少妇粗大呻吟视频| 久久久精品94久久精品| 国产熟女欧美一区二区| 久久久久久免费高清国产稀缺| 亚洲综合色网址| 国产精品九九99| 精品少妇一区二区三区视频日本电影| 久久久久久亚洲精品国产蜜桃av| 十八禁网站网址无遮挡| 水蜜桃什么品种好| 两人在一起打扑克的视频| 男女床上黄色一级片免费看| 曰老女人黄片| 性少妇av在线| 午夜福利,免费看| 男人爽女人下面视频在线观看| 一区二区av电影网| 人妻一区二区av| 手机成人av网站| 操美女的视频在线观看| 人人妻人人爽人人添夜夜欢视频| 亚洲一区中文字幕在线| 国产一区二区在线观看av| 欧美黄色片欧美黄色片| 男人爽女人下面视频在线观看| 亚洲国产av影院在线观看| 高清视频免费观看一区二区| 国产精品久久久久久精品电影小说| 菩萨蛮人人尽说江南好唐韦庄| 亚洲成av片中文字幕在线观看| 老司机在亚洲福利影院| 一级毛片 在线播放| www.av在线官网国产|