• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Template Matching Based Feature Extraction for Activity Recognition

    2022-08-24 12:57:58MuhammadHameedSiddiqiHelalAlshammariAmjadAliMadallahAlruwailiYousefAlhwaitiSaadAlanaziandKamruzzaman
    Computers Materials&Continua 2022年7期

    Muhammad Hameed Siddiqi, Helal Alshammari, Amjad Ali, Madallah Alruwaili,Yousef Alhwaiti, Saad Alanaziand M.M.Kamruzzaman

    1College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, 2014, Kingdom of Saudi Arabia

    2Department of Computer Science, COMSATS University Islamabad, Lahore Campus, Pakistan

    Abstract: Human activity recognition (HAR) can play a vital role in the monitoring of human activities, particularly for healthcare conscious individuals.The accuracy of HAR systems is completely reliant on the extraction of prominent features.Existing methods find it very challenging to extract optimal features due to the dynamic nature of activities, thereby reducing recognition performance.In this paper, we propose a robust feature extraction method for HAR systems based on template matching.Essentially, in this method, we want to associate a template of an activity frame or sub-frame comprising the corresponding silhouette.In this regard, the template is placed on the frame pixels to calculate the equivalent number of pixels in the template correspondent those in the frame.This process is replicated for the whole frame, and the pixel is directed to the optimum match.The best count is estimated to be the pixel where the silhouette (provided via the template)presented inside the frame.In this way, the feature vector is generated.After frame, and the pixel is directed to the optimum match.The best count is to label the incoming activity.We utilized different publicly available standard datasets for experiments.The proposed method achieved the best accuracy against existing state-of-the-art systems.

    Keywords: Activity recognition; feature extraction; template matching; video surveillance

    1 Introduction

    Human activity recognition has a significant role in many applications such as tele medicine and healthcare, neuroscience, and crime detection.Most of these applications need additional grades of independence like rotation or orientation, scale or size and viewpoint distortions.Rotation might be felt by spinning the template or by utilizing Arctic coordinates; scale invariance might be attained using templates of various size.Having additional parameters of attention infers that the accumulator space becomes bigger; its dimensions rise through one for every extra parameter of attention.Positioninvariant template matching infers a 2D parameter space; while, the enlargement of scale and positioninvariant template matching needs 3D parameter space [1].

    Human activity recognition (HAR) systems try to automatically recognize and examine human activities by acquiring data from different sensors [2].HAR is frequently associated to the procedure of finding and naming actions using sensory annotations [3].Generally, a human activity states the movement of one or many parts of the human body, which might be static or composed of numerous primitive actions accomplished in some successive order.Hence, HAR should permit classification the same activity with the similar label even when accomplished by various persons under various dynamic [2].

    There are various types of audio and video sensors that can be employed in HAR systems.However, most of them have their own limitations.In audio sensors-based data collection, we may lose the data because of utilizing GPRS to transmit the data.This is one the main disadvantages of the audio-based data collection.Therefore, in this work, we will be using video-sensor (such as 2D RGB camera).HAR system has three basic stages.In the first stage, the noise and environmental distortion will be diminished from the video frame.Furthermore, in this stage, we also segment the human body.In the second stage, we extract the best and informative features from the segmented body.While, in the last stage, a classifier is employed to categorize the incoming activities as shown in Fig.1.

    Figure 1: General flow diagram of a HAR system

    Commonly, classification has two types: First is the frame-based classification; while, second is sequence-based classification.In frame-based classification, only the present frame is employed with or without a standard frame in order to categorize the human actions from the arriving videos.On the other hand, in the sequence-based classification, the symmetrical movement of the feature pixels is considered among the present frame and the preliminary frame.Therefore, the frame-based classification does not have the ability in such domains in order to classify human activities; hence, the concentration of this work is the sequence-based classification [4].

    Accordingly, some latest works have been developed for the sequence-based HAR systems that showed significant performance in various dynamic scenarios.A state-of-the-art system was proposed by [5-8] that is based on the extraction of the individual persons’scene from the sequence of frames.Then, 3D convolutional neural network was utilized in order to detect and classify the corresponding activities of every sequence of frames.Activity-based video summarization is accomplished by saving every person’s activity at every time of the incoming video.Similarly, another sequence-based HAR system is proposed by [9] for the identification of the human in healthcare domains.This system takes video frames of COVID-19 patients, then finds for a match inside the grip on frames.In this system, the Gabor filter is utilized for feature extraction where the personal sample generation formula along with Gabor filter is utilized on input frame in order to collect the optimum and non-redundant Gabor features.Further, deep learning models are employed for matching the human activities with input frame.Furthermore, a robust sequence-based HAR system was proposed by [10] that was assessed on Weizmann and KTH actions datasets.In the pre-processing step of this system, the authors extracted the initial frames from input videos and resized.Then, frame by frame, the region of interest has been considered by employing Blob detection technique and tracing is done with the help of Kalman filter.Furthermore, an ensembled method (which is a group of various techniques such as bidimensionalempiricalmode decomposition, scale invariant feature transform, and wavelet transform)was employed for feature extraction, which extracts the features from moving object.Similarly, this method was also utilized on pre-processed frames in order to extract the best features from multiscaled frames.Finally, convolution neural network was employed for activity classification.Most of these systems suffer from their own limitations such as the degradation of accuracy in dynamic and naturalistic environments.

    Therefore, in this work, we have proposed an adoptive feature extraction method.Essentially in this method, we want to associate a template of an activity frame which will be the template like subframe which comprises the silhouette, we are going to search.Therefore, we focus the template on the frame pixels and calculate the equivalent number of pixels in the template correspondent those in the frame.This process is replicated for the whole frame, and the pixel that directed to the optimum match, the best count is estimated to be the pixel where the silhouette (provide via the template)presented inside the frame.For the experiments,we utilized various publicly available standard datasets such as Weizmann dataset [11], KTH action dataset [12], UCF sports dataset [13], and IXMAS action dataset [14] respectively.The proposed technique showed best performance against existing works.

    The remaining article is ordered as: Section 2 provides some recent literature review about sequence-based human activity classification systems.The detailed description on the proposed feature extraction is presented in Section3.The detailed description on the proposed feature Section4.The Section 5 describes the experimental setup.While, in Section 6, the results along with the discussion are explained.Lastly, in Section 7, the proposed HAR system will be summarized along with little future directions.

    2 Related Work

    Human activity states the movement of one or many parts of the human body, which might be static or composed of numerous primitive actions accomplished in some successive order.There lots of state-of-the-art methods have been proposed for HAR systems.However, most of them their own limitations.The authors of [15] developed a state-of-the-art system that is based on the architecture of deep learning and V4 inception in order to classify the incoming activities.However, deep learning lacks mutual intelligence, which makes the corresponding systems flimsy and the errors might be very large if the errors are made [16].Moreover, due to the larger number of layers, the step time of Inception-v4 is suggestively slower in practice [17].

    Similarly, an HAR system was proposed by [18] that is based on dissimilarity in body shape,which has been divided into five parts that associate to five fractional occupancy regions.For every frame, the region ratios have been calculated that further be employed for classification purpose.For classification, they utilized the advantages of AdaBoost algorithm that has the greater acumen capacity.However, AdaBoost algorithm cannot be equivalent since every predictor might only be trained after the preceding one has been trained and assessed [19].A novel ensembled model was proposed by [20] for Har systems, where they utilized multimodal sensor dataset.They proposed a new data preprocessing method in order to permit context reliant feature extraction from the corresponding dataset to be employed through various machine learning techniques such as linear discriminant,decision trees, kNN, cubic SVM, DNN, and bagged tree.However, every of these algorithms has its own limitations, for instance, kNN, SVM and DNN are frame-based classifiers that do not have the ability to accurately recognize the human activities from incoming sequences of video frames [21].

    A new HAR approach was introduced by [22] which is based on entropy-skewness and dimension reduction technique in order to get the condensed features.These features are then transformed into a codebook through serial-based fusion.In order to select the prominent and best features, a genetic algorithm is applied on the created feature codebooks, and for classification, a multi-class SVM has been employed.However, the well-known limitation of the genetic algorithm is that it does not guarantee any variety amongst the attained solutions [23].Moreover, SVM does not have the capability to correctly classify the human activities from incoming sequences of video frames [21].A naturalistic HAR system was proposed by [24] for which the human behavior is demonstrated as a stochastic sequence of activities.Activities are presented through a feature vector including both route data such as position and velocity, and a group of local movement descriptors.Activities are classified through probabilistic search of frames feature records on behalf of formerly seen activities.Hidden Markov Models (HMM) was employed for activity classification from incoming videos.However, the local descriptors have one of the main limitations, means that due to this algorithm the results might not be directly transferred to pixel descriptors which cannot be further utilized for classification [25].

    A motion-based feature extraction was proposed by [26] for HAR systems.They employed the context information from various resources to enhance the recognition.So, for that purpose, they presented the scene context features which presents the situation of the subject at various levels.Then for classification, the structure of deep neural network was utilized in order to get the higher-level presentation of human actions, which further combined with context features and motion features.However, deep neural network has major limitations such as short transparency and interpretability,and requires huge amount of data [27].Moreover, the motion features are very scant if human or background comprise non-discriminative features, and sometimes, the extracted features are defectiveand vanish in succeeding frames[28].Avery recent systemwas proposed by [29] that is based on various machine learning techniques such as Spatio-temporal interest point, histogram orient gradient, Gabor filter, Harris filter coupled with support vector machine, and they claimed best accuracy.However, the aforementioned techniques have major limitations such as the high-frequency response of Gabor filter produces ring effect closer to the edges which may degrade the accuracy [30].Moreover, Harris filter requires much time for feature extraction and space to store them, which might not be suitable for naturalistic domains [31].

    On the other hand, an automatic sequence-based HAR system was proposed by [32], which is based on group features along with high associations into category feature vectors.Then every action is classified through the amalgamation of Gaussian mixture models.However, Gaussian mixture model is a frame-based classifier which does not has the ability to accurately classify video-based activities.Another sequence-based HAR system was designed by [33] that was based on the neural network.The corresponding networkswere created the features database of various activities that were extracted and selected from sequence of frames.Finally, multi-layer feed forward perceptron network was utilized used in order to classify the incoming activities.However, neural network is a vector-based classifier that has low performance against sequence of frames [21].Similarly, a multi-viewpoint HAR systems was proposed by [34] that was based on two-stream convolutional neural networks integrated with temporal pooling scheme (that builds non-direct feature subspace depictions.However, their accuracy was very low in naturalistic domains.Moreover, temporal pooling scheme receive the shortcomings in performance generalization as described in [35] that clearly make the benefit of trained features over handmade ones [36].

    A multimodal scheme was proposed for human action recognition [37].This system was based on ascribing importance to the semanticmaterial of label texts instead of just mapping them into numbers.After this step, they modelled the learning framework that reinforces the video description with additional semantic language management and allows the proposed model to the activity recognition without additional required parameters.However, semantic information has some major issues like dimension detonation, data sparseness, incomplete generalization capacity [38].

    Accordingly, this work presents an accurate, robust and dynamic feature extraction method that has the ability to extract the best features from the sequence of video frames.In this method, we want to associate a template of an activity frame which will be the template like sub-frame which comprises the silhouette, we are going to search.Therefore, we focus the template on the frame pixels and calculate the equivalent number of pixels in the template correspondent those in the frame.This process is replicated for the whole frame, and the pixel that directed to the optimum match, the best count is estimated to be the pixel where the silhouette (provide via the template) presented inside the frame.By this way, the feature vector is generated.After feature extraction, the hidden Markov model(HMM) has been utilized in order to label the incoming activities.

    3 Proposed Feature Extraction Method

    In a typical human activity recognition system, the accuracy is completely relying on the feature extraction module.Therefore, we proposed a robust and naturalistic method for feature extraction module.In this method, we want to associate a template of an activity frame which will be the template like sub-frame which comprises the corresponding silhouette.Therefore, we focus the template on the frame pixels and calculate the equivalent number of pixels in the template corresponding to those in the frame.This process is replicated for the whole frame, and the pixel that is directed to the optimum match, the best count is estimated to be the pixel where the silhouette (provided via the template) is inside the frame.

    Generally, template matching might be explained as an algorithm of parameter calculation.The parameters describe the template location in the image, which might be defined as a distinct function Fi, jthat accepts the values in a frame such as the coordinates of the pixels like(i, j)∈S.For instance,a set points of 3×3 template may be defined as S={(0, 0, 0) (0, 0, 1) (0, 1, 0) (0, 1, 1) (1, 0, 0)(1, 0, 1) (1, 1, 0) (1, 1, 1)}.

    Let assume that every pixel in the activity frameImi, jis disturbed by the noise of additive Gaussian,and the corresponding noise is the mean of zero and the unidentified standard deviation that is represented by σ.Hence, the probability at a pixels’template positioned at the coordinates(x, y)ties the equivalent pixel at location(i, j)∈S that is shown by the general distribution

    where σ indicates the Gaussian distribution and π is the ratio between the edges and diameter of the image area.Meanwhile, the noise that affect every pixel is autonomous, the probability of the template at location(x, j)is the fused probability of every pixel that is covered by the template, such as

    Put Eq.(1), then we have

    wherekrepresents the number of points in the corresponding template, which is known as the likelihood function.Commonly, for simpler analysis, this function is expressed in the form of logarithmic.It should be noticed that the scale of the logarithm function does not modify the location of the maximum likelihood.Hence, the updated likelihood function under the logarithm is shown as shown below

    To select the parameter which enlarges the likelihood function, we need to estimate the maximum likelihood.For instance, the location enlarges the rate of modification of the objective function.

    Hence, the aforementioned equations also provide the solution of the minimization issue, which is given as

    Here, the estimation of maximum likelihood is equal to picking the location of the template which diminishes the shaped errors.The location where the utmost matches of the frame template is the projected location of the template inside the frame.Hence, if the solution of maximum likelihood has been selected based on the measurement of the matching under the criteria of squared error.This indicates that the result attained via template matching is optimum for frames that are crooked through Gaussian noise.It should be noted that practically assessed noise might be presumed to be the Gaussian noise based on the recommendation of the algorithm of the central limit, though many frames seem to deny this presumption.Alternatively, other errors criteria like the complete difference,instead of the squared difference.

    The alternative criteria of the squared error can be derived by substituting Eq.(7), which can be written as:

    The final part of the Eq.(8) does not rely on the location of the template(x, y).Intrinsically, it is continuous and might not be diminished.Hence, the optimal in Eq.(8) might be gained through minimizing.

    If the initial term

    is almost continuous, then the rest of the terms give a quantity of the likeness among the template and frame.Specifically, we might enlarge the cross correlation among the frame and template.Hence, the best location might be calculated as

    But, the term of square in Eq.(10) may be changed with location; so, the defined match through Eq.(11) might be poor.Similarly, the variety of the cross-correlation is reliant on the template size,which means that under various environmental conditions, it does not vary.Hence, it is more feasible to utilize either Eqs.(7) or (9) in implementation.

    On the other hand, in order to normalize the cross-correlation, Eq.(8) can be defined as below

    Accordingly, the first part is consistent, and hence, the optimal value might be attained as

    Generally, it is feasible to stabilize the window for every activity frame against the template.So,

    whereImx, yis the average of the pixelsImi+x, j+y, which is utilized for points inside the window (such as(i, j)∈S) and F indicates the is the average of the pixels in the corresponding template.Likewise,normalized cross-correlation is presented by Eq.(14), which does not modify the location of the optimal and provides a clarification as the vector of cross-correlation is normalized.Hence,

    If the activity frame and the corresponding template are binary, then such type of combination for template matching will be more beneficial, which might present the regions in the frame or it may comprise the edges.The overall flowchart of the proposed approach is presented in Fig.2.

    Figure 2: The flowchart of the proposed feature extraction approach

    4 Utilized Action Datasets

    The proposed feature extraction technique has been tested and validated on four publicly available standard action datasets such as Weizmann dataset, KTH action dataset, UCF sports dataset, and IXMAS action dataset respectively.Every action dataset is explained as below:

    4.1 Weizmann Dataset

    In this dataset, there are ten various activities which are performed by nine different subjects.The corresponding activities are skip, bend, walk, run, side changing, place jumping, forward jumping, one hand waving (Wave-1) and two hand waving (Wave-2) respectively.The dataset has total 90 activity clips having approximately 15 frames/activity.In order to normalize the entire frames of the dataset,we resized them to 280×340.

    4.2 KTH Action Dataset

    This dataset was created by 25 subjects who performed total six activities such as walk, boxing,run, clapping, jogging, and waving in various dynamic distinctive situations.This dataset was created under the setting of static camera against consistent background.The dataset has total 2391 sequences under the size of were taken with a frame size 280×320.

    4.3 UCF Sports Dataset

    This dataset contains 182 videos of total that were assessed through n-fold cross validation scheme from television channels.This dataset was created from various sports persons who were performing different sport matches.Moreover, the entire activities were collected under the settings of static camera.Some of the classes have high intra-class resemblances.There is total nine activities such as diving, run, lifting, skating, golf swimming, kick, walk, baseball swimming, and horse back riding.Each activity frame has a size 280×320.

    4.4 IXMAS (INRIA Xmas Motion Acquisition Sequences) Action Dataset

    In this dataset, there were total thirteen activities that were performed by eleven subjects.Each actor selected a free angle and location.For each subject, there were corresponding silhouettes in this dataset.We have chosen eight activity classes such as cross arm, walk, turn around, punch, wave, sit down, kick, and get up.This dataset has a view-invariant HAR where the size of each activity frame is size 280×320 (for our experiments).This dataset suffers from high occlusion which may reduce the performance of the proposed approach; therefore, we employed one of our previous methods [39] to normalize the occlusion concern.

    5 Experiments Setup

    The proposed method was assessed and validated against the following set of experiments.

    5.1 First Experiment

    This experiment presents the accuracy of the HAR system under the presence of the proposed feature extraction technique.So, for that purpose, we performed four sub-experiments against each dataset in order to show the significance and robustness of the proposed technique.

    5.2 Second Experiment

    This experiment indicates the role and importance of the designed approach in a typical HAR system.So, we utilized an inclusive set of sub-experiments for such persistence.For these experiments,we employed various state-of-the-art feature extraction methods instead of using the proposed technique.

    5.3 Third Experiment

    Finally, in this experiment, we compared the accuracy of the proposed method against state-ofthe-art systems.

    6 Results and Discussions

    6.1 First Experiment

    In this sub-experiment, we presented the performance of the proposed feature extraction technique against each dataset.For reach dataset,we utilizedn-fold cross validation structure,which means that every activity is utilized for training and testing respectively.The overall result of the proposed method is shown in Tab.1 (Weizmann dataset), Tab.2 (KTH action dataset), Tab.3 (UCF dataset),and Tab.4 (IXMAS dataset) respectively.

    Table 1: Analysis of the proposed approach on Weizmann dataset

    Table 2: Analysis of the proposed approach on KTH action dataset

    It should be noted from Tabs.1-4 that the common HAR system along with the proposed feature extraction method achieved accuracy on every dataset.From these results, we observed that the proposed method is robust, which means the proposed feature extraction method did not achieve best accuracy only on one dataset but also showed significant performances on other datasets respectively.This is because the averaging intrinsic in the proposed feature extraction method is the reduction of the vulnerability to noise and the maximization stage diminishes defenselessness to occlusion.

    Table 3: Analysis of the proposed approach on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    Table 4: Analysis of the proposed approach on IXMAS dataset, where CrA is Cross Arm, SiD is Sit Down, GeU is Get Up, TuA is Turn Around

    6.2 Second Experiment

    For this experiment, we performed a group of sub-experiments in order to show the performance of the proposed HAR system.The entire sub-experiments were performed on every dataset under the absence of the proposed feature extraction method.For these sub-experiments, we utilized recent well-known feature extraction techniques such as wavelet transform [4], Curvelet transform [40], local binary pattern (LBP) [41], local directional pattern (LDP) [42], and stepwise linear discriminant analysis (SWLDA) [43] respectively.The overall results of the sub-experiments are presented in Tabs.5-24 against Weizmann dataset, KTH dataset, UCF dataset, and IXMAS dataset of various activities.

    Table 5:Analysis of a common HAR system along with existing wavelet transform(without employing the proposed approach) on Weizmann dataset

    Table 6: Analysis of a common HAR system along with existing Curvelet transform (without employing the proposed approach) on Weizmann dataset

    Table 7: Analysis of a common HAR system along with existing local binary patter (LBP) (without employing the proposed approach) on Weizmann dataset

    Table 7: Continued

    Table 8: Analysis of a common HAR system along with existing local directional pattern (LDP)(without employing the proposed approach) on Weizmann dataset

    Table 9: Analysis of a common HAR system along with existing stepwise linear discriminant analysis(SWLDA) (without employing the proposed approach) on Weizmann dataset

    Table 10: Analysis of a common HAR system along with existing wavelet transform (without employing the proposed approach) on KTH action dataset

    Table 11: Analysis of a common HAR system along with existing Curvelet transform (without employing the proposed approach) on KTH action dataset

    Table 12: Analysis of a common HAR system along with existing local binary pattern (LBP) (without employing the proposed approach) on KTH action dataset

    Table 13: Analysis of a common HAR system along with existing local directional pattern (LDP)(without employing the proposed approach) on KTH action dataset

    Table 14: Analysis of a common HAR system along with existing stepwise linear discriminant analysis(SWLDA) (without employing the proposed approach) on KTH action dataset

    Table 15: Analysis of a common HAR system along with existing wavelet transform (without employing the proposed approach) on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    Table 16: Analysis of a common HAR system along with existing Curvelet transform (without employing the proposed approach) on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    Table 17: Analysis of a common HAR system along with existing local binary patter (LBP) (without employing the proposed approach) on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    As can be seen from Tabs.5-24 that under the absence of the proposed approach (like feature extraction technique), the HAR system did not achieved best accuracy.This is because the inattentiveness to noise and occlusion are the main benefits of the proposed feature extraction technique.Noise may happen in any frame of the incoming video.Similarly, there might be low noise in digital photographs; however, in image processing it is made inferior through edge detection by the quality of variation procedures.Furthermore, shapes might simply be obstructed or hidden, for instance, a person may walk behind a streetlamp, or illumination may be one of reasons to create occlusion.The averaging intrinsic in the proposed feature extraction method is the reduction of the vulnerability to noise and the maximization stage diminishes defenselessness to occlusion.

    Table 18: Analysis of a common HAR system along with existing local directional patter (LDP)(without employing the proposed approach) on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    Table 19: Analysis of a common HAR system along with existing stepwise linear discriminant analysis (SWLDA) (without employing the proposed approach) on UCF dataset, where GoS is Golf Swimming, HoBR is Horse Back Riding, and BaS is Baseball Swimming

    Table 20: Analysis of a common HAR system along with existing wavelet transform (without employing the proposed approach) on IXMAS dataset, where CrA is Cross Arm, SiD is Sit Down,GeU is Get Up, TuA is Turn Around

    Table 21: Analysis of a common HAR system along with existing Curvelet transform (without employing the proposed approach) on IXMAS dataset, where CrA is Cross Arm, SiD is Sit Down,GeU is Get Up, TuA is Turn Around

    Table 22: Analysis of a common HAR system along with existing local binary pattern (LBP) (without employing the proposed approach) on IXMAS dataset, where CrA is Cross Arm, SiD is Sit Down,GeU is Get Up, TuA is Turn Around

    Table 22: Continued

    Table 23: Analysis of a common HAR system along with existing local directional pattern (LDP)(without employing the proposed approach) on IXMAS dataset, where CrA is Cross Arm, SiD is Sit Down, GeU is Get Up, TuA is Turn Around

    Table 24: Analysis of a common HAR system along with existing stepwise linear discriminant analysis(SWLDA) (without employing the proposed approach) on IXMAS dataset, where CrA is Cross Arm,SiD is Sit Down, GeU is Get Up, TuA is Turn Around

    6.3 Third Experiment

    Finally, in this group of experiments, we have compared the recognition rate of the proposed approach against latest HAR systems.For some system,we have borrowed their implementation code;while, for the remaining system, we have presented their accuracies as described in their respective articles.The entire systems were implemented under the exact settings as indicated in their respective articles.For comparison, we also utilized, UCF50 dataset [44] and HMDB51 dataset [45].The comparison results are accordingly presented in Tab.25.

    Table 25: Performance of the proposed approach against recent HAR systems

    It is vibrant fromTab.25 that the proposed approach achieved best weighted average classification accuracy against state-of-the-art works.The reason is that, the proposed technique has the capacity to extract the prominent features from the action frames under the presence of occlusion, illumination and background disorder and scale changes.Moreover, the proposed approach extracts the best features from various resources such as shapes, textures, and colors in order to build the feature vector that will be input for a classifier.

    7 Conclusions

    Human activity recognition (HAR) has a fascinating role in our daily life.HAR can be applied for healthcare domains to check the patients’daily routines.Also, HAR has a significant role in other applications such as crime control, sports, defense etc.There are many resources for HAR systems.Among them, video-camera is one of the best candidates for HAR systems.The accuracy of such systems completely depends upon the extraction and selection of the best features from the activity frames.Accordingly, in this work, we have proposed a new feature extraction technique that is based on template matching.In the proposed approach, we matched a template of an image which will be the template like sub-frame which comprises the silhouette.Therefore, we focus the template on the frame pixels and calculate the equivalent number of pixels in the template correspondent those in the frame.The proposed approach was assessed against four publicly available standard datasets of activities, which sowed showed the best performance against existing recent HAR systems.The averaging intrinsic in the proposed approach is the reduction of the vulnerability to noise and the maximization stage diminishes defenselessness to occlusion.Moreover, the proposed algorithm has the capacity to extract the prominent features from the activity frames under the presence of occlusion, illumination and background disorder and scale changes.Also, the proposed approach extracts the best features from various resources such as shapes, textures, and colors for building the feature vector that will be input for a classifier.

    In the future, we will implement and deploy the proposed HAR system under the presence of the proposed feature extraction in healthcare, which will facilitate the physicians to remotely check the daily exercises of the patients through which they might easily recommend the corresponding recommendations for the patients.This approach may also help the patients sufficiently improve the quality of their lives in healthcare and telemedicine.

    Funding Statement:The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia for funding this work through the Project Number“375213500”.Also,the authors would like to extend their sincere appreciation to the central laboratory at Jouf University to support this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    中文天堂在线官网| 成人无遮挡网站| 能在线免费观看的黄片| 中文字幕人妻熟人妻熟丝袜美| 成年女人在线观看亚洲视频 | 国产精品国产三级专区第一集| 一级毛片 在线播放| 日本爱情动作片www.在线观看| 老司机影院成人| 干丝袜人妻中文字幕| 精品国内亚洲2022精品成人| 色网站视频免费| 亚洲成人精品中文字幕电影| 一级毛片电影观看| 成人性生交大片免费视频hd| 国产 一区 欧美 日韩| 人妻制服诱惑在线中文字幕| 久久精品国产亚洲av涩爱| 国产成人a区在线观看| 日韩,欧美,国产一区二区三区| 国产一级毛片七仙女欲春2| 高清日韩中文字幕在线| 69av精品久久久久久| 亚洲精品第二区| 精品人妻一区二区三区麻豆| 成人高潮视频无遮挡免费网站| 久久热精品热| 国产亚洲午夜精品一区二区久久 | 亚洲av在线观看美女高潮| 国产伦一二天堂av在线观看| 亚洲真实伦在线观看| 777米奇影视久久| 男女视频在线观看网站免费| 成人亚洲精品av一区二区| 成人一区二区视频在线观看| 你懂的网址亚洲精品在线观看| 看非洲黑人一级黄片| 听说在线观看完整版免费高清| 国产成人aa在线观看| 老女人水多毛片| 黄色配什么色好看| 国产精品国产三级专区第一集| 有码 亚洲区| 亚洲精品国产av成人精品| 免费黄网站久久成人精品| 中国美白少妇内射xxxbb| 亚洲精品亚洲一区二区| 亚洲av电影在线观看一区二区三区 | 亚洲欧美清纯卡通| 国产高清三级在线| 亚洲国产精品成人久久小说| 极品少妇高潮喷水抽搐| 亚洲成色77777| 精品人妻视频免费看| 亚洲精品乱久久久久久| 久久久久久伊人网av| 淫秽高清视频在线观看| 国产精品一区www在线观看| 极品少妇高潮喷水抽搐| 免费观看在线日韩| 亚洲熟妇中文字幕五十中出| 久久久久精品性色| 日韩一区二区视频免费看| 你懂的网址亚洲精品在线观看| 美女国产视频在线观看| 91精品伊人久久大香线蕉| 国产精品国产三级国产专区5o| 欧美潮喷喷水| 成人毛片a级毛片在线播放| 美女大奶头视频| 男女边吃奶边做爰视频| 久久精品国产亚洲网站| 特级一级黄色大片| 婷婷色综合大香蕉| 美女内射精品一级片tv| 日本三级黄在线观看| 九草在线视频观看| 亚洲欧美成人综合另类久久久| 九色成人免费人妻av| 99久久精品国产国产毛片| 精品久久久久久久久久久久久| 成年版毛片免费区| 最近最新中文字幕免费大全7| 日韩av在线大香蕉| av线在线观看网站| 欧美日韩一区二区视频在线观看视频在线 | 高清视频免费观看一区二区 | 亚洲av不卡在线观看| 国产美女午夜福利| 18禁在线无遮挡免费观看视频| 男插女下体视频免费在线播放| 亚洲精品乱码久久久v下载方式| 99久久九九国产精品国产免费| 搡女人真爽免费视频火全软件| 日日摸夜夜添夜夜爱| 国产单亲对白刺激| 伦精品一区二区三区| 日日啪夜夜撸| 免费观看无遮挡的男女| 国产精品久久久久久久久免| 看十八女毛片水多多多| 精品久久久久久电影网| 丰满少妇做爰视频| 成人综合一区亚洲| 国产乱人视频| 欧美一区二区亚洲| 精品99又大又爽又粗少妇毛片| 天天一区二区日本电影三级| 久久99热这里只有精品18| 欧美成人a在线观看| 丝瓜视频免费看黄片| 夫妻性生交免费视频一级片| 黑人高潮一二区| 2021少妇久久久久久久久久久| 亚洲人成网站高清观看| 亚洲激情五月婷婷啪啪| 免费大片18禁| 丝瓜视频免费看黄片| 精品国产一区二区三区久久久樱花 | 极品少妇高潮喷水抽搐| 久久热精品热| 成人午夜高清在线视频| 乱码一卡2卡4卡精品| 99视频精品全部免费 在线| 国产成人91sexporn| 人人妻人人澡欧美一区二区| 免费观看无遮挡的男女| 黄色欧美视频在线观看| 日日啪夜夜撸| 精品久久久久久久末码| 久久精品熟女亚洲av麻豆精品 | 精品久久久精品久久久| 久久久欧美国产精品| av在线观看视频网站免费| 91在线精品国自产拍蜜月| 亚洲人成网站在线播| 在线a可以看的网站| 最近最新中文字幕免费大全7| 国产精品嫩草影院av在线观看| 成人一区二区视频在线观看| 日韩中字成人| 成人无遮挡网站| 成人综合一区亚洲| av免费观看日本| 久久韩国三级中文字幕| 男女那种视频在线观看| 日产精品乱码卡一卡2卡三| 欧美日韩视频高清一区二区三区二| 久久草成人影院| 黄色日韩在线| 亚洲av男天堂| 午夜激情福利司机影院| 三级男女做爰猛烈吃奶摸视频| 丰满人妻一区二区三区视频av| 男人爽女人下面视频在线观看| 80岁老熟妇乱子伦牲交| 少妇丰满av| 插阴视频在线观看视频| 国产成人午夜福利电影在线观看| 日韩不卡一区二区三区视频在线| 国产黄色视频一区二区在线观看| 黄色一级大片看看| 国产成人精品婷婷| 午夜精品国产一区二区电影 | 亚洲精品亚洲一区二区| 久久精品久久精品一区二区三区| 国产色婷婷99| av国产免费在线观看| 午夜福利在线观看吧| 亚洲av成人精品一区久久| 22中文网久久字幕| 卡戴珊不雅视频在线播放| 国产高清三级在线| 久久国产乱子免费精品| 一本一本综合久久| 人人妻人人看人人澡| 日本色播在线视频| 日韩亚洲欧美综合| 国产精品一区二区三区四区免费观看| 亚洲欧洲日产国产| 亚洲人成网站在线播| 寂寞人妻少妇视频99o| 国产精品爽爽va在线观看网站| 三级毛片av免费| 中文字幕久久专区| 国产日韩欧美在线精品| 精品国内亚洲2022精品成人| 国产精品综合久久久久久久免费| 亚洲内射少妇av| 亚洲av免费高清在线观看| 久久精品熟女亚洲av麻豆精品 | 亚洲欧美成人综合另类久久久| 成人无遮挡网站| 久久久久久久亚洲中文字幕| 国产永久视频网站| 有码 亚洲区| 在线免费观看不下载黄p国产| 99久久精品一区二区三区| 国产永久视频网站| 最近视频中文字幕2019在线8| 韩国高清视频一区二区三区| 赤兔流量卡办理| 国产极品天堂在线| 免费看美女性在线毛片视频| 三级国产精品片| 九九久久精品国产亚洲av麻豆| 国产免费视频播放在线视频 | 精品人妻视频免费看| 搡老妇女老女人老熟妇| 内地一区二区视频在线| 欧美最新免费一区二区三区| 黄片wwwwww| 精品久久久精品久久久| 日本午夜av视频| 又爽又黄无遮挡网站| 亚洲人成网站在线播| 亚洲精华国产精华液的使用体验| 国内精品一区二区在线观看| 欧美xxⅹ黑人| av国产免费在线观看| 麻豆国产97在线/欧美| 一区二区三区乱码不卡18| 特大巨黑吊av在线直播| 亚洲性久久影院| 2021天堂中文幕一二区在线观| 男女边吃奶边做爰视频| 99热6这里只有精品| 观看免费一级毛片| 久久久久久久大尺度免费视频| 亚洲欧美日韩东京热| 日本wwww免费看| 日本一二三区视频观看| 精品国内亚洲2022精品成人| 精品人妻偷拍中文字幕| 亚洲精品国产成人久久av| 美女高潮的动态| 大香蕉久久网| 亚洲欧洲国产日韩| 国产精品久久视频播放| 午夜久久久久精精品| 永久免费av网站大全| 激情五月婷婷亚洲| 男人狂女人下面高潮的视频| 波野结衣二区三区在线| 久久久久久久久久黄片| 日韩 亚洲 欧美在线| 亚洲三级黄色毛片| 日产精品乱码卡一卡2卡三| 深爱激情五月婷婷| 久久人人爽人人片av| 免费人成在线观看视频色| 精品酒店卫生间| 3wmmmm亚洲av在线观看| 日本免费a在线| 国产在线男女| 又爽又黄无遮挡网站| 国产欧美日韩精品一区二区| 三级国产精品欧美在线观看| 亚州av有码| 丰满少妇做爰视频| 亚洲精品国产av蜜桃| av福利片在线观看| 菩萨蛮人人尽说江南好唐韦庄| 亚洲精品影视一区二区三区av| 精品亚洲乱码少妇综合久久| 免费av不卡在线播放| 草草在线视频免费看| 成人午夜高清在线视频| 中文字幕制服av| 久久久久免费精品人妻一区二区| 午夜福利网站1000一区二区三区| 亚洲自偷自拍三级| 一本久久精品| 极品少妇高潮喷水抽搐| 熟女电影av网| 久久这里只有精品中国| 69av精品久久久久久| 亚洲av不卡在线观看| 免费看光身美女| 五月伊人婷婷丁香| 成人午夜高清在线视频| 嘟嘟电影网在线观看| 熟妇人妻久久中文字幕3abv| 国产精品福利在线免费观看| 免费看美女性在线毛片视频| 国产成人aa在线观看| 亚洲精品一二三| 国产欧美另类精品又又久久亚洲欧美| 免费看日本二区| 久久热精品热| 成年人午夜在线观看视频 | 日本午夜av视频| 国产黄色视频一区二区在线观看| 国产熟女欧美一区二区| 少妇猛男粗大的猛烈进出视频 | 中文欧美无线码| 2021少妇久久久久久久久久久| 午夜免费男女啪啪视频观看| 国产国拍精品亚洲av在线观看| 久久久久久久久久久丰满| 中文欧美无线码| 水蜜桃什么品种好| 国产一级毛片七仙女欲春2| 丝袜美腿在线中文| av在线天堂中文字幕| 嫩草影院精品99| 国产色婷婷99| 国产伦精品一区二区三区视频9| 天堂俺去俺来也www色官网 | 最近手机中文字幕大全| 欧美激情久久久久久爽电影| 亚洲最大成人av| 国产综合懂色| 搡女人真爽免费视频火全软件| 久久精品久久久久久久性| 天天躁夜夜躁狠狠久久av| 国内精品宾馆在线| 国内精品美女久久久久久| 性插视频无遮挡在线免费观看| 在线播放无遮挡| 精品久久国产蜜桃| 成人午夜精彩视频在线观看| 男女国产视频网站| 人人妻人人看人人澡| 高清av免费在线| 成人美女网站在线观看视频| 亚洲av成人av| 免费少妇av软件| 免费电影在线观看免费观看| 国产综合懂色| 亚洲伊人久久精品综合| 色尼玛亚洲综合影院| 亚洲18禁久久av| 久久久久久久大尺度免费视频| 欧美日韩在线观看h| 一级黄片播放器| 久久热精品热| 亚洲精品成人av观看孕妇| 麻豆久久精品国产亚洲av| 2021少妇久久久久久久久久久| 黄色配什么色好看| 在线观看免费高清a一片| 久久鲁丝午夜福利片| 偷拍熟女少妇极品色| 日韩在线高清观看一区二区三区| 欧美性感艳星| 国产午夜精品论理片| 亚洲精品国产成人久久av| 国产伦精品一区二区三区四那| 高清av免费在线| 日韩视频在线欧美| 色哟哟·www| 国产色婷婷99| 精品人妻偷拍中文字幕| 高清日韩中文字幕在线| 国产高清三级在线| 久久精品夜夜夜夜夜久久蜜豆| 性插视频无遮挡在线免费观看| 人妻制服诱惑在线中文字幕| 色播亚洲综合网| 久久精品国产亚洲av涩爱| 国产黄a三级三级三级人| 国产片特级美女逼逼视频| 黄色日韩在线| 国语对白做爰xxxⅹ性视频网站| 亚洲人成网站在线观看播放| 一级二级三级毛片免费看| 免费少妇av软件| 街头女战士在线观看网站| 亚洲成人久久爱视频| 少妇高潮的动态图| 噜噜噜噜噜久久久久久91| 国产色爽女视频免费观看| 亚洲欧美精品专区久久| 99热这里只有是精品在线观看| 亚洲国产高清在线一区二区三| 嫩草影院精品99| 免费人成在线观看视频色| 精品久久久久久电影网| av在线老鸭窝| av国产免费在线观看| 晚上一个人看的免费电影| 亚洲欧美精品自产自拍| 欧美性猛交╳xxx乱大交人| 日日摸夜夜添夜夜添av毛片| 国产精品国产三级国产av玫瑰| 男女那种视频在线观看| 国产精品久久久久久精品电影| 最后的刺客免费高清国语| 国产伦在线观看视频一区| 精品午夜福利在线看| 国产成人精品一,二区| 少妇被粗大猛烈的视频| 国产精品国产三级国产专区5o| 男人舔奶头视频| 美女cb高潮喷水在线观看| 欧美高清性xxxxhd video| 亚洲人成网站在线观看播放| 久久精品夜夜夜夜夜久久蜜豆| 高清毛片免费看| 亚洲综合色惰| 中文字幕久久专区| 天堂√8在线中文| 亚洲图色成人| 男的添女的下面高潮视频| 少妇人妻精品综合一区二区| 全区人妻精品视频| 毛片女人毛片| av天堂中文字幕网| 国产精品精品国产色婷婷| 内射极品少妇av片p| 国产成年人精品一区二区| 亚洲精品成人av观看孕妇| 欧美精品一区二区大全| 免费观看的影片在线观看| 人人妻人人看人人澡| 久久久久久久久久黄片| 欧美高清性xxxxhd video| 免费av毛片视频| 久热久热在线精品观看| 最近中文字幕2019免费版| 亚洲精品国产av成人精品| www.av在线官网国产| 国产精品久久视频播放| 看十八女毛片水多多多| 亚洲国产欧美人成| 观看免费一级毛片| 成人国产麻豆网| 国产综合懂色| 婷婷色综合大香蕉| 国产综合懂色| www.色视频.com| 最近2019中文字幕mv第一页| 久久99蜜桃精品久久| 国内精品宾馆在线| 午夜日本视频在线| 国产老妇伦熟女老妇高清| 亚洲精品自拍成人| 成人av在线播放网站| 中文字幕制服av| 全区人妻精品视频| 视频中文字幕在线观看| 色5月婷婷丁香| 高清午夜精品一区二区三区| 国产片特级美女逼逼视频| 久久久欧美国产精品| 一个人免费在线观看电影| www.av在线官网国产| 噜噜噜噜噜久久久久久91| 日本熟妇午夜| 视频中文字幕在线观看| 久久精品国产鲁丝片午夜精品| 日本午夜av视频| 又爽又黄无遮挡网站| 亚洲美女视频黄频| 久久久久久久国产电影| 亚洲激情五月婷婷啪啪| www.av在线官网国产| 国产精品日韩av在线免费观看| 91精品一卡2卡3卡4卡| 男女边吃奶边做爰视频| 乱码一卡2卡4卡精品| 一级片'在线观看视频| 乱码一卡2卡4卡精品| 色哟哟·www| 精品久久久噜噜| 激情 狠狠 欧美| 中文字幕久久专区| 少妇的逼好多水| 麻豆成人av视频| 国产国拍精品亚洲av在线观看| 国产亚洲精品av在线| 激情 狠狠 欧美| 大又大粗又爽又黄少妇毛片口| 亚洲精品国产成人久久av| 国产免费又黄又爽又色| 国产精品精品国产色婷婷| 夜夜爽夜夜爽视频| 亚洲伊人久久精品综合| 国产中年淑女户外野战色| 99久久人妻综合| av又黄又爽大尺度在线免费看| 国产亚洲av片在线观看秒播厂 | 欧美丝袜亚洲另类| 久久精品国产自在天天线| 国产中年淑女户外野战色| 久久精品久久久久久噜噜老黄| 狂野欧美白嫩少妇大欣赏| 人体艺术视频欧美日本| 成人漫画全彩无遮挡| 欧美成人午夜免费资源| 久久久亚洲精品成人影院| 日本欧美国产在线视频| 亚洲激情五月婷婷啪啪| 三级毛片av免费| 日韩伦理黄色片| 男人狂女人下面高潮的视频| 免费观看无遮挡的男女| 18+在线观看网站| 精品午夜福利在线看| 亚洲精华国产精华液的使用体验| 国产视频首页在线观看| 九草在线视频观看| 嫩草影院入口| 91久久精品电影网| 丰满少妇做爰视频| 午夜福利在线在线| 免费看av在线观看网站| 日韩大片免费观看网站| 中文字幕人妻熟人妻熟丝袜美| 亚洲熟女精品中文字幕| 观看免费一级毛片| 在线免费观看的www视频| 看黄色毛片网站| 亚洲熟妇中文字幕五十中出| 伊人久久精品亚洲午夜| 亚洲欧美日韩卡通动漫| 天天一区二区日本电影三级| 国产精品久久视频播放| 中文字幕av成人在线电影| 日韩欧美精品免费久久| 99视频精品全部免费 在线| 免费在线观看成人毛片| 免费看a级黄色片| 国产成人a∨麻豆精品| 日韩亚洲欧美综合| av福利片在线观看| 午夜激情欧美在线| 亚洲欧美精品专区久久| 又黄又爽又刺激的免费视频.| 偷拍熟女少妇极品色| 一级毛片黄色毛片免费观看视频| 久久人人爽人人片av| 亚洲伊人久久精品综合| 99热这里只有精品一区| 国产精品三级大全| 久久这里有精品视频免费| 在线免费观看的www视频| 欧美潮喷喷水| 久久韩国三级中文字幕| 熟女电影av网| 日韩av在线大香蕉| 大陆偷拍与自拍| 久久久久九九精品影院| 街头女战士在线观看网站| 亚洲一区高清亚洲精品| 精品午夜福利在线看| 免费观看精品视频网站| 一二三四中文在线观看免费高清| 黄色日韩在线| 久久久久性生活片| 国产亚洲5aaaaa淫片| 直男gayav资源| 久久99热这里只有精品18| 亚洲精华国产精华液的使用体验| 激情五月婷婷亚洲| 免费少妇av软件| 搡老乐熟女国产| 色网站视频免费| 日本欧美国产在线视频| 欧美xxⅹ黑人| 欧美极品一区二区三区四区| 国产精品一及| 不卡视频在线观看欧美| 一个人观看的视频www高清免费观看| 成人毛片a级毛片在线播放| 麻豆精品久久久久久蜜桃| 在线免费观看的www视频| 免费观看精品视频网站| av卡一久久| 成人亚洲欧美一区二区av| 成年免费大片在线观看| 国内揄拍国产精品人妻在线| 成人一区二区视频在线观看| 久久精品综合一区二区三区| 美女脱内裤让男人舔精品视频| 国产精品熟女久久久久浪| 91av网一区二区| 高清在线视频一区二区三区| 舔av片在线| 日韩国内少妇激情av| 国产人妻一区二区三区在| 边亲边吃奶的免费视频| 水蜜桃什么品种好| 精品国产三级普通话版| 亚洲18禁久久av| 国产午夜精品一二区理论片| 日韩成人av中文字幕在线观看| 亚洲欧美一区二区三区黑人 | 看免费成人av毛片| 国产高清有码在线观看视频| 欧美精品一区二区大全| 啦啦啦啦在线视频资源| 日韩av免费高清视频| 97超碰精品成人国产| 亚洲四区av| ponron亚洲| 国产精品蜜桃在线观看| 91av网一区二区| 十八禁国产超污无遮挡网站| 久久97久久精品| 久久鲁丝午夜福利片| 国产中年淑女户外野战色| 日本午夜av视频| 免费看不卡的av| 色播亚洲综合网| 2021天堂中文幕一二区在线观| 中文字幕制服av| 最近视频中文字幕2019在线8| 小蜜桃在线观看免费完整版高清| 69av精品久久久久久| 日产精品乱码卡一卡2卡三| 欧美成人午夜免费资源| 亚洲av一区综合| 伦理电影大哥的女人|