• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Identification and Classification of Crowd Activities

    2022-08-24 12:59:12ManarElshahawyAhmedAseeriShakerElSappaghHassanSolimanMohammedElmogyandMervatAbuElkheir
    Computers Materials&Continua 2022年7期

    Manar Elshahawy, Ahmed O.Aseeri, Shaker El-Sappagh, Hassan Soliman,Mohammed Elmogyand Mervat Abu-Elkheir

    1Department of Information Technology, Faculty of Computers and Information, Mansoura University, Egypt

    2Department of Computer Science, College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj, 11942, Saudi Arabia

    3Department of Information Systems, Faculty of Computers and Artificial Intelligence, Benha University, Egypt

    4Faculty of Computer Science and Engineering, Galala University, 435611, Suez, Egypt

    5Department of Computer Science, Faculty of Media Engineering and Technology, German University, Egypt

    Abstract: The identification and classification of collective people’s activities are gaining momentum as significant themes in machine learning, with many potential applications emerging.The need for representation of collective human behavior is especially crucial in applications such as assessing security conditions and preventing crowd congestion.This paper investigates the capability of deep neural network (DNN) algorithms to achieve our carefully engineered pipeline for crowd analysis.It includes three principal stages that cover crowd analysis challenges.First, individual’s detection is represented using the You Only Look Once (YOLO) model for human detection and Kalman filter for multiple human tracking; Second, the density map and crowd counting of a certain location are generated using bounding boxes from a human detector; and Finally, in order to classify normal or abnormal crowds, individual activities are identified with pose estimation.The proposed system successfully achieves designing an effective collective representation of the crowd given the individuals in addition to introducing a significant change of crowd in terms of activities change.Experimental results on MOT20 and SDHA datasets demonstrate that the proposed system is robust and efficient.The framework achieves an improved performance of recognition and detection people with a mean average precision of 99.0%, a real-time speed of 0.6ms non-maximum suppression (NMS) per image for the SDHA dataset,and 95.3% mean average precision for MOT20 with 1.5ms NMS per image.

    Keywords: Crowd analysis; individual detection; You Only Look Once(YOLO); multiple object tracking; kalman filter; pose estimation

    1 Introduction

    Crowd analysis has recently gained much attention for its broad range of applications usage,including video surveillance.It has been an active research topic in machine learning (ML) and computer vision using continuous videos [1].Crowded situations may originate from numerous scenarios, posing safety and security concerns.The crowd size and dynamics of crowds affect the safety of the people.Hence, further studies are needed on crowds and individuals in the crowd for surveillance and safe pre-control [2].Because crowds usually have so many individuals accumulated with frequent and substantial movements.Many existing detections and tracking technologies, which are only applicable to sparse crowd scenes, do not work well in crowded scenes.The number of surveillance cameras installed all over the world has increased unprecedentedly.This makes it easier for us to identify and classify a wide area of crowds.Yet, with limited human resources, many video images simultaneously are not sufficient to analyze them.So, there is a need to have an automated way to monitor the population and classify them.The development of an intelligent surveillance system is one of the key applications of crowd analysis.In general, an analysis of a crowd may be conducted at three levels [3].First, macroscopic levels where a mass of people’s global motions and behaviors are the focus of interest.So, we can consider the whole crowd at once with no concern for the movements of individuals.Second, microscopic levels where the research deals with the behaviors and activities of any individuals in the crowd [4].Finally, mesoscopic levels are based on combinations of intrinsic characteristics of microscopic and macroscopic levels.

    This paper incorporates deep learning (DL) algorithms into our framework to improve identification accuracy, allowing better data to be derived from videos and analyzed per crowd [5-7].So,analyzing individual behavior in crowded scenes is one of the most important and challenging areas in current research.The main steps of our framework: Firstly, we start with identifying the crowd and use object detection techniques to analyze crowds at microscopic levels [8].Individuals in a crowd can move in the same direction or randomly.This is one of the challenges of crowd recognition and identification with high density in the MOT20 dataset in unconstrained environments.Therefore, we want to create a framework for crowd density representation using You Only Look Once (YOLO)by detecting individuals who appear in the video [9].YOLO is used as a real-time object detection algorithm.It has the advantage of being much faster than other networks and still maintains accuracy[10,11].Detecting individuals is the most important step in our research.Our model is completed with the result of the YOLO algorithm.The second most difficult challenge is multiple object tracking,which localizes detected persons along with the video [12].Thirdly, the computation of calculating the total number of individuals is considered the third challenging problem [13].Challenges appear due to occlusions, constant displacement of the person, varied viewpoints and behavior, varying levels of illumination, and the allocation of pixels per person decreases in high density.Finally, the proposed article focuses on observing abnormal human activities.A crowd analysis method has been completed using human pose estimation to provide an automated approach for detecting abnormal crowds.The major contributions of this paper can be highlighted as follows:

    ?A novel framework is proposed for DL model-based multi-scale feature extraction from video frames.

    ?A modified YOLO is used for a more accurate and fast human detection method in an unconstrained environment.

    ?A collective representation of the crowd is designed using a density map, given the individuals’positions.

    ?Crowd counting is produced for high density, based on human detection.

    ?For more precise crowd classification, individual activities and several algorithms for human tracking are used.

    The remainder of the article is arranged as follows.Section 2 summarizes the proposed study’s most applicable methodologies in the related work section.Our proposed DL architecture is introduced in Section 3, especially employing human activities for crowd classification.A detailed experimental result and datasets of our work are followed in Section 4.Finally, we conclude and suggest possible future research directions in Section 5.

    2 Related Work

    There is much research in crowd analysis.Each of them had his standard or real dataset to solve specific problems, so most studies are not comparable to others.Previous work on crowd analysis is summarized in Tab.1.

    Table 1: A summary of some current related work studies

    Table 1: Continued

    Direkoglu et al.[14] proposed an optical flow-based event feature extraction of the abnormal crowd.The angle difference between optical flow vectors was computed in each pixel’s current and previous frames.Then multiplication of the angle difference with the optical flow magnitude in the current frame was used to remove noise.In the end, one class SVM was used to identify normal crowd behavior, and abnormal crowd activity was observed when a test sample deviated significantly from normal behavior.Using the UMN dataset, the optimal window size was 141, but in PETS2009, the optimal window size was 71.This paper needs to apply motion optical flow for a precise system.

    Zheng et al.[15] used WISDM v1.1 (daily activity knowledge collected by a smartphone in an exceedingly laboratory, with a rate of 20 cycles/s), WISDM v2.0 (daily activity knowledge collected by a smartphone during an uncontrolled system with a rate of 20 cycles/s) and Skoda (manipulative gestures performed in an automotive maintenance situation, with the rate of 98 cycles/s).WISDM is collected from 51 test subjects performing 18 activities for 3 min each.In this paper, the knowledge about eight activities was accelerated.The participants completed every activity for approximately 10 min except for jumping (approximately 1.7min).Classifiers were separately trained and evaluated for each participant, which meant extra execution time.

    Yan et al.[16] used the ShanghaiTech dataset, which was split into Part A and Part B, with 482 and 716 images.PGCNet was represented for crowd counting.The PGC block was built by combining the features before/after the PGC module.The problem occurs at low picture resolution.

    Asghari et al.[17] utilized two datasets collected using 32 and 30 sensors from 2 different houses for five months.Each house had one person.The model hierarchically comprised a series of HMM,hereby developed to recognize the start of the activity, analyze the ongoing activity, and analyze the ongoing activity, identify the end and class of activities.Additional work needs to be done to improve recognition of the other class as the most troublesome class of activity.

    Castanon et al.[18] used a crowd activity data set of 5,668 examples from 17 activities of NIST Extended Video Actions (ActEV) challenge for vehicles or people’s activities and associated objects Out The Window (OTW).The VIRAT data set was employed for training and validation,which includes 1,404 train Set activities and 1203 Validation Set activities.There are inconsistencies hence the cause of these inconsistencies needs to be identified, and the OTW dataset expanded for addressing them.

    Vahora et al.[19] proposed multi-layer deep architecture as human action-pose level CNN followed by recurrent neural network (RNN) model.LSTM and GRU models were employed to tackle the long-range dependency of the simple RNN model.Framework started with spatial feature extraction of person for action-pose classification using 16 classes of activities.The result showed that the LSTM model with Adam optimizer, learning rate 0,001, with N=64 number of units in LSTM cell and the GRU model with RMSprop optimizer, learning rate 0,001, with N=32 number of GRU units boosts performance compared to other variations.

    Chauhan et al.[20] used gradient histogram and the Blob analytical method to extract the feature and recognize human movement within the video frame.The SVM then categorizes activities (normal(77.91%) or abnormal (82.99%) and types of activities such as Hectic (88.36%), Fight (86.87%), panic(79.1%), then crowd types were given.

    Guo et al.[21] applied two resident datasets (Tulum and Cairo) provided by the Center for Advanced Studies in Adaptive Systems (CASAS).They evaluated his technique by comparing it with several conventional classifiers.Fourteen activity categories from 36 sensors were collected in Tulum2010 at 98 days measurement time, and Cairo had been collected from 32 sensors and 57 measurement days in 13 categories of activity.Two married residents who performed routine daily activities were applied to a home with practical experiments.Event feature extraction is based on separating activities that occur in the same space but at different times.

    The WMNL2016 dataset collected from the Wireless & Mobile Network Laboratory was used by Chen et al.[22].In WMNL2016, 40 state-of-the-art sensors were put in a six-area apartment, while one voluntary worker carried out 12 months’activities.In 35 or more training days, the number of activities recognized was 12, which was used to tailor the Service for the elderly for individuals in the aged care sector in a smart home.In order to predict and recognize abnormal activity recognition, this article should employ deep learning and some other master learning technologies.

    In crowded scenes, Abdullah et al.[12] presented framework counting, monitoring, and detecting the behavior of several persons with gradient motion descriptors and Spatio-temporal entropy classifier.The first dataset, PETS2009, was a data set that includes the Sample 1 data set used to count people in a low-density crowd.The Sample 2 data set was used to detect and track humans in medium-density crowds.The Sample 3 data set was used for tracking and estimating the number of individuals in a large crowd.The second dataset was UMN, consisting of three separate scenes (one indoors and two outdoors) to detect abnormal crowd behavior.The lawn scenario was one of the outdoor scenes, which consisted of 1453 frames and the Plaza scene was the second outdoor scene with three scenarios with 2142 frames.Six indoor scenes with 4144 frames were presented.Occlusion and single scene detection were the problems in this framework.Different scenes are considered, like sports, combat, traffic, and action scenes.

    Not only the object in the image is classified, but the object is also localized to consist of an object detection model [23,24].Two stages had been employed in earlier years to address this problem: first,sliding windows of different sizes were used to extract different areas in the image.Second, it applied the classification problem to determine what class the objects belong to.The problem with these approaches is that they need a great deal of calculation and are divided into several stages.This makes it difficult for the system to be optimized in terms of speed.Using the YOLO algorithm, our framework reconstructs object detection from image pixels to bounding box coordinates and class probabilities as a single regression issue [25].The system provides prediction vectors that match each object in the image.The idea of the YOLO system is that it calculates all the features of the image.Besides, it predicts all the items simultaneously instead of iterating the process of classifying the different areas in the image.

    3 The Proposed Crowd Analysis Technique

    This paper introduces a crowd analysis framework in which the majority of problems are solved.In particular, it is especially difficult to detect each item individually if two or more individuals move close together; the human detector output is a merged result.Another important concern is irregular human distribution when the density differs in a video.These systems are faced with a non-uniform arrangement of different individuals that are close together.This is what is known as clutter.To deal with these challenges, the proposed approach is described in Fig.1.The initial phase, employing the ML model for human detection, is crowd detection.This work requires quick human detection to be carried out in real-time.The network is trained with frames in which the crowd is included.After the training stage is properly done, people’s activities in frames are classified to see their effect on the crowd’s classification.

    Figure 1: The flow of the crowd analysis framework

    Our approach includes three primary stages of crowd analysis: individual detection, crowd representation, and classification.They are discussed in the following activities using video data(MOT20 and SDHA datasets).The first step in human detection is preprocessing, which focuses on formulating datasets to obtain videos containing a crowd using the YOLO model and tracking algorithm.In the second stage, identification and crowd map representation are applied in selected videos that contain a crowd.They are utilized to identify individuals as a part of the crowd and calculate the total number of individuals in each frame.In the third stage, the individual activities are classified to identify normal/abnormal crowds.Finally, the training model assigns individuals to the crowd to detect if the crowd changed using the number of individuals and activities.

    3.1 Object Detection of Individuals

    3.1.1 Preprocessing of Human Detection

    The purpose of the preprocessing phase on the human detector is to provide an acceptable input in real-time for the model’s suitability.Datasets are divided randomly into training, validation, and test sets using handcrafted methods.The extraction of input video feed frames is necessary.These frames are resized and smoothed out individually.The original frame is normalized, which improves the detecting speed without losing any precious frame data.This is a crucial method to ensure all individuals are labeled, especially those with the high-density crowd, to improve performance accuracy.It is done for the computing parameters needed and for further implementation and detection.The processed stage output is given into the YOLO model.After integrating the anchor box size of the detection model for classification purposes, preprocessing procedures would be finalized.

    3.1.2 YOLOv5-S Model Architecture

    YOLO is a neural network method that recognizes and detects different objects(here-individuals).It has grown in popularity over the above-mentioned detection technology due to its improved performance and uses only 7MB of memory.Human detection is done in YOLO as a regression issue and offers the class probability for the observed items as well as bounding boxes.YOLOv5-S is one of the various versions of YOLO, which consists of three main components that distinguish it from other algorithms [26].Firstly, the backbone model is a CNN that generates important features from given images using Cross Stage Partial network (CSP) and Focus interlaced sampling splicing structure as the backbone.CSPNET deals with the issue of repeated gradient information in large-scale backbones[27].It integrated gradient changes into the feature map.It reduces computations by reducing model parameters and FLOPS (floating-point operations per second), which improves inference speed and accuracy and reduces model size.Secondly, the Neck model is used to form feature pyramids.Feature pyramids aid models in achieving good object scaling generalization.It aids in identifying the same object in various sizes and scales.

    Path Aggregation Network (PANET) [28] and Spatial Pyramid Pooling (SPP) [29] are used as the neck model of YOLOv5’s.It boosts information flow and improves the use of exact location signals at lower layers, improving the object location precision.This version uses a subsampling layer that reduces the spatial scale of the convolved features.The computation power needed to process the data should be decreased by reducing dimensionality.Thirdly, YOLOv5’s head, the YOLO layer, is used to do the final detection part.It generates a multi-scale prediction of three feature maps, allowing the model to handle small, medium, and large objects.Generalized Intersection overUnion (GIoU-loss) is used to maximize the overlap area of the ground truth and predicted bounding box of detected object[30].Backbone, Neck, and head models of YOLOv5’s have been displayed in Fig.2.The stochastic gradient descent (SGD) was employed [31].It starts with an initial learning rate using ADAM as the adaptive learning rate [32].SoftMax is used in the final layer of the network, representing a single class’s utilization.The results of the detection stage are transformed to the training stage, as in Fig.3.

    Figure 2: The architectural description of YOLOv5’s

    Figure 3: The flowchart for the training of YOLOv5’s

    These results contain the bounding coordinate vector [pc,bx,by,bw,bh,c] of each detection with a probability of person classpc, bounding box coordinate of the detected objectbx,by,bw,bhandcis the class.Then annotations of each image are saved in the form of a.txt files, with each line describing a bounding box.After that, YOLOv5 is used for the training stage [33].We have provided a model with a special extra layer to improve efficiency once the Yolov5 original model has been implemented.We introduced an additional scale including certain convolutional layers with activation functions to improve object detection, especially occlusion problems.

    3.1.3 Crowd Counting

    Crowd counting is determined using detection-based approaches, which are utilized to determine each individual’s representation in the crowd.These methods consider the crowd as an independent collection of people who can detect individuals by generating bounding boxes and then counting people with bounding boxes.Crowd size is computed from the total number of bounding boxes in each frame [34,35].

    3.2 Multiple Individual Tracking

    Detection is followed by tracking individuals using bounding boxes and assigning a unique identifier to each one [36-39].Using SORT tracking as a framework for Kalman filtering and Hungarian optimization to track individuals listed in Algorithm 1.A recursive estimator, the Kalman filter at timet+1,is used to compute the estimation for the current state at timet.All that is required is the estimated state from the previous time step and the current measurement.As a result, the human’s location can be maintained even if the human’s vision is blocked.By assessing whether an individual in the current frame is the same individual detected in the previous frames.Furthermore, the Hungarian algorithm assigns a unique ID number to a target individual.Each individual in a frame is described by the following Eq.(1):

    The target bounding box’s horizontal and vertical positions (i.e., the centroid) are represented by (u,v), whereassspecifies the scale (area) andrdenotes their aspect ratios.Horizontal position,vertical position, and bounding box centroid are predicted by Kalman filter foru′,v′and s′.The existing bounding box is updated with the newly observed information when a recognized individual is associated with a new observation.This is determined using the Kalman filter framework’s estimated speed and acceleration parameters.If the expected IDs of the query individual significantly change from the new observation, the Kalman filter employs the same state that it was predicted.Otherwise,the weights of the corrections are evenly distributed between the new observation and the Kalman filter prediction.Following the detection and tracking phase, we define the matrixDtat Eq.(2), which includes the location ofndetected individuals in the frame carrier grid using bounding box priorP,for each input frameIw×hat timet:

    Algorithm 1: Multiple human tracking using Kalman Filter and Hungarian algorithms Input: Frames, Kalman Filter, Hungarian algorithms, Tracking Bounding boxes (t-1), Detection Bounding boxes (t)Output: Tracked indivimages/BZ_809_675_1670_676_1672.pngiduals and Updated Kalman Filter 1.For each video do 2.Create KalmanFilterTracker;3.For each frame in the video do 4.Prepare the frame data;5.Predict object locations;6.calculate IOU using tracking and detection bounding boxes;7.match predicted with actual using the Hungarian algorithm;8.If not equal Then 9.Update KalmanFilterTracker;10.End If 11.End for 12.simple output trackers;13.End for

    3.3 Crowd Representation

    To represent the crowd map, the hue color range on the HSV channel was previously defined[40,41].Abounding box is generated for each detected individual, which is utilized to calculate the individual’s location in the heatmap.In general, the redder grids, the more individual bounding boxes are likely to intersect with one another.In addition to people’s basic movement and tracking data,analyzing crowd density might be more valuable.So, risky zones might be identified or redesigned to secure the environment.

    3.4 Crowd Classification

    After the crowd is represented, we need to analyze human activities to know how they can reflect the impact of their changes on crowds.Crowd control can be improved as a result of a greater understanding of human activities and crowd dynamics using multiple human tracking listed in Algorithm 2.

    3.4.1 Top-Down Pose Estimation

    Pose estimation is utilized to identify and predict human activities such as sitting, standing,lying down, or punching or jumping [42-44].Top-down approaches for pose estimation run a person detector first and then estimate body joints.There is a lot of overlap between individuals, different individual scales, and occlusion in our datasets in crowd frames.Therefore, a lot of difficulties have been solved with top-down techniques.The critical points in an individual’s body include 18 joints such as the elbow, knees, and wrists are localized using the TensorRT pose method using ResNet backbone.A person’s body representation is built using the position of human body parts from visual input data,as shown in Fig.4.As a result, a person’s body modeling is essential for estimating human position.It represents extracted features and key points from visual input data after detection and tracking using the YOLO technique and Kalman filter.

    Figure 4: Overall pipeline to estimate human pose for activity recognition

    3.4.2 Crowd Labeling

    After that, there are nine activities.They are labeled as normal and abnormal activities.In the case of abnormal individual activities, the crowd is classified as abnormal using the handcrafted method.

    4 Experimental Results

    4.1 Experimental Platform

    The model’s training was conducted using MOT20 and SDHA datasets in a free Integrated Development Environment (IDE) called Google Collaboratory (Google Colab).Tesla K80 is utilized from Google Colab, which is Google provides to support research and learning about ML.The preinstalled libraries of Google Colab are pretty and allow the DL model to gain accuracy and learn from additional data sets.

    Algorithm 2: Human Pose Estimation then crowd classification Input: Frames, trt-pose estimation, Bounding boxes, tracking output Output: Activity and crowd classification 1.For each video do 2.Create object detection and Tracking algorithms;3.For each frame in the video do 4.Prepare the frame data;5.Get the joints’positions by trt-pose;6.If joints position is empty Then 7.fill with joints’relative position in previous frame;8.End If 9.Use DNN to activity classifier;10.If activity is abnormal Then 11.Assign abnormal label of crowd;12.Else 13.Assign normal label of crowd;14.End If 15.Track each person;16.End for 17.simple output classification;18.End for

    4.2 Dataset Description

    4.2.1 MOT20

    Eight new sequences in unconstrained environments were chosen from three crowded situations for the MOT20 benchmark [45].As part of prior challenges, all sequences were carefully collected and chosen for human detection and tracking.This benchmark focuses on scenarios with high pedestrian density, which can exceed 246 pedestrians per frame.In addition to indoor and outdoor locations, the sequences contain both day and night shots.

    4.2.2 SDHA

    It contains videos of human-to-human encounters, including handshakes, pointing, hugging,pushing, and punching [46,47].Time intervals and bounding boxes are provided for these interactions as ground truth labels.All twenty video sequences are roughly one minute in length for training, and there are three videos for testing.There is at least one execution per interaction in each video, resulting in an average of eight human activities in each movie.Aperson’s height in the video is around 200 pixels.The videos are 720 * 480 at 30 frames per second.Videos are divided into two groups.A parking lot is the setting for the first two sets of videos.Backgrounds are generally static, and camera jitter is minimal in set 1.Set 2 (i.e., the last ten sequences) was shot on a breezy lawn.Backgrounds are shifting slightly(trees, for example), and there are greater camera jitters in these videos as a result.

    4.3 Model Evaluation Indicators

    Precision (P), also known as precision rate Eq.(3), is a measure used in this paper to show how well the model predicted the individuals.The number of actually detected people divided by the total of truly detected individuals and undetected people in each frame is known as the recall rate (R) in Eq.(4).The harmonic mean of precision and recall is denoted as the F1 score, which is calculated using Eq.(5).False positives (FP), true positives (TP), and false negatives (FN) are all used to calculate precision and recall as given below:

    Furthermore, mAP (mean average precision) is computed using Eq.(6).In the detection model,GIoU quantifies the overlap between the bounding box of predicted and the real object’s ground truth bounding box.Every value of the IoU threshold provides a different mAP.Therefore, this value must be specified.If IoU is compared to a specific threshold, the detection may be valid or wrong.To test individual detection model performance using a series of experiments, we test the performance of the trained system using GIoU thresholds of 0.5 and 0.5:0.95.

    4.4 Results

    4.4.1 Training the YOLOv5

    In this stage, learnable weights and biases are assigned to objects in preprocessing frames.Then,the YOLO algorithm is applied with initial parameters.For example, see Tab.2.

    Table 2: Training parameter settings

    4.4.2 Summary of Results

    The trained model of object-detection is executed rapidly and makes it possible to utilize in realtime with a prediction in a fraction of a second.Tab.3 shows the results of those metrics for MOT20 and SDHA datasets using YOLOv5s using 100 epochs.As a result of a different resolution, SDHA is executed with 640 image size, but MOT20 is 1280 image size.

    Table 3: Performance of the model YOLOv5s for MOT20 and SDHA datasets for 100 epochs

    4.4.3 SDHA Metrics Representation

    Using the mAP is a good measure of the neural network’s sensitivity in Fig.5 to describe the research outcomes.At network size, the Precision-Recall Curve (PR-curve) in Fig.6 is executed, and evaluations of it at the GIoU threshold extend from 0.5 to 0.95.We can observe that the model performed well overall, with network size 224 and GIoU threshold 0.5 having the highest mAP value of .99 percent.

    Figure 5:The result of mean average precision at two thresholds, precision and recall for SDHA dataset

    4.4.4 MOT20 Metrics Representation

    The model performed mAP well overall at MOT20 in Fig.7, with network size 283 and GIoU threshold 0.5 having the highest mAP value of 0.953 percent.At network size, the Precision-Recall Curve (PR-curve) in Fig.8 is executed, and evaluations of it at the GIoU threshold extend from 0.5 to 0.95.

    Figure 6: F1 score and PR curve of YOLOv5 for SDHA dataset

    Figure 7: The result of mean average precision at two thresholds, precision and recall for MOT20 dataset

    Figure 8: F1 score and PR curve of YOLOv5 for MOT20 dataset

    5 Discussion

    This model investigated the capability of deep learning to complete crowd analysis with detected bounding boxes.Although the utilized datasets have different densities and are unlabeled, the object detection algorithm worked well.After that, multiple human tracking and human activity recognition algorithms are utilized to improve YOLO to detect more individuals, especially in more crowded places.Finally, the proposed detection model utilizing MOT20 is more accurate than other algorithms listed in Tab.4.We report the results on average accuracy (AP), multi-object detection accuracy(MODA),recall,and precision.We notice that the YOLOv5 detector specializes in pedestrian detection in crowded scenes, and it achieves high precision and is easily implemented.Another advantage is reducing the amount of computation due to one stage detector.We need to study the labeling technique for crowded places for further improvement since the amount of training data directly determines detection performance.Besides that, we need to add more activities to be estimated for more accuracy.

    Table 4: Detection evaluation on MOT20 dataset

    6 Conclusion

    This paper presents a deep learning-based approach for object detection, multiple human tracking, and crowd classification using human activities in videos.YOLO technique for object detection was applied in 2 different datasets with different densities to ensure performance experimentally.Experimental results reveal that the YOLOv5 algorithm is superior through its speed, storage, and object detection accuracy.The bounding box from the previous phase was transferred in multiple human tracking and activity recognition stages.TheKalman filter algorithm is employed as a tracking algorithm.It was applied under many items such as activity changing and movement of objects, appearance changes, size changes, and slight changes in MOT20 and SDHA datasets.Besides, the presented system can efficiently recognize anomalous human activity from the crowd utilizing pose estimation, which significantly improves the system’s accuracy and efficiency.Unusual Crowd Activity Detection can be used in various public places to provide early warning of crime and improve crowd management.In future research, we plan to study the inclusion of more abnormal activities.The ability to train deeper convolutional layers on additional datasets can also help the model perform better.

    Funding Statement:The authors received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    熟女人妻精品中文字幕| 中文天堂在线官网| 亚洲一级一片aⅴ在线观看| 69人妻影院| 麻豆精品久久久久久蜜桃| 日本一本二区三区精品| 国产精品久久久久久久电影| 日韩欧美三级三区| 国内精品宾馆在线| 国产亚洲av片在线观看秒播厂 | 777米奇影视久久| 一个人观看的视频www高清免费观看| 国产单亲对白刺激| 久久久精品免费免费高清| 三级国产精品欧美在线观看| 久久久国产一区二区| 美女cb高潮喷水在线观看| 内地一区二区视频在线| 少妇高潮的动态图| 免费播放大片免费观看视频在线观看| 亚洲欧美日韩卡通动漫| 国产91av在线免费观看| 成年版毛片免费区| 色综合亚洲欧美另类图片| 亚洲欧美日韩卡通动漫| 久久久久久久久久久丰满| 欧美不卡视频在线免费观看| 九九久久精品国产亚洲av麻豆| 日本爱情动作片www.在线观看| 国产视频内射| 日韩国内少妇激情av| av播播在线观看一区| 我要看日韩黄色一级片| 一级av片app| 97超碰精品成人国产| 国内少妇人妻偷人精品xxx网站| 亚洲18禁久久av| 久久久久九九精品影院| 成人美女网站在线观看视频| 国产爱豆传媒在线观看| 我要看日韩黄色一级片| 成人亚洲精品av一区二区| 久久鲁丝午夜福利片| 精品人妻视频免费看| 美女xxoo啪啪120秒动态图| 韩国av在线不卡| 久久久欧美国产精品| 久久精品人妻少妇| 91久久精品国产一区二区三区| 国产精品一区www在线观看| 亚洲三级黄色毛片| 神马国产精品三级电影在线观看| 国产精品女同一区二区软件| 欧美极品一区二区三区四区| 欧美日韩综合久久久久久| 日韩制服骚丝袜av| 国内精品宾馆在线| 午夜免费男女啪啪视频观看| 只有这里有精品99| 日本爱情动作片www.在线观看| 成人二区视频| 人人妻人人澡人人爽人人夜夜 | 久久精品久久久久久久性| 久久亚洲国产成人精品v| 欧美三级亚洲精品| 国产老妇女一区| 国产av国产精品国产| 69人妻影院| 久久99热这里只频精品6学生| 午夜亚洲福利在线播放| 午夜福利在线观看吧| 久久久久九九精品影院| 大香蕉97超碰在线| 精品久久久久久久久av| 亚洲熟妇中文字幕五十中出| 日日干狠狠操夜夜爽| 好男人在线观看高清免费视频| 久久久久久久国产电影| 午夜免费男女啪啪视频观看| 三级男女做爰猛烈吃奶摸视频| 麻豆久久精品国产亚洲av| 国产一级毛片七仙女欲春2| 色网站视频免费| 久久久久久九九精品二区国产| 99热网站在线观看| 好男人视频免费观看在线| a级毛色黄片| 观看免费一级毛片| 精品国产一区二区三区久久久樱花 | 深夜a级毛片| www.色视频.com| av在线天堂中文字幕| 国产在视频线精品| 99久久人妻综合| 久久久久国产网址| 人妻夜夜爽99麻豆av| 日韩欧美三级三区| 亚洲va在线va天堂va国产| 亚洲国产精品成人久久小说| 女人被狂操c到高潮| 国产精品不卡视频一区二区| 欧美xxⅹ黑人| 99热网站在线观看| 一边亲一边摸免费视频| 2021少妇久久久久久久久久久| 又黄又爽又刺激的免费视频.| 国产一区二区三区综合在线观看 | 七月丁香在线播放| 国产精品日韩av在线免费观看| 校园人妻丝袜中文字幕| 国产片特级美女逼逼视频| 欧美成人精品欧美一级黄| 夜夜爽夜夜爽视频| 网址你懂的国产日韩在线| 国产亚洲5aaaaa淫片| 美女黄网站色视频| 女人十人毛片免费观看3o分钟| 特大巨黑吊av在线直播| 亚洲av日韩在线播放| 丰满乱子伦码专区| 亚洲久久久久久中文字幕| 精品国产露脸久久av麻豆 | 女的被弄到高潮叫床怎么办| 天堂中文最新版在线下载 | 女人十人毛片免费观看3o分钟| 色尼玛亚洲综合影院| 国产成人精品久久久久久| 色综合站精品国产| 内地一区二区视频在线| 中文字幕av成人在线电影| 听说在线观看完整版免费高清| 国产精品久久久久久久久免| 国产午夜福利久久久久久| 97在线视频观看| 免费看av在线观看网站| 精品熟女少妇av免费看| 综合色丁香网| 一级爰片在线观看| 亚洲欧美日韩无卡精品| 亚洲四区av| 18+在线观看网站| 国产午夜精品久久久久久一区二区三区| 久久97久久精品| 尤物成人国产欧美一区二区三区| 亚洲真实伦在线观看| 日本爱情动作片www.在线观看| 日日啪夜夜爽| 国产精品一区二区性色av| 青春草亚洲视频在线观看| 亚洲av电影不卡..在线观看| 亚洲国产高清在线一区二区三| 色综合站精品国产| 国产在视频线在精品| 人妻夜夜爽99麻豆av| 99九九线精品视频在线观看视频| 久久久久久久国产电影| 日韩一本色道免费dvd| 一个人免费在线观看电影| 国产在视频线在精品| 午夜福利高清视频| 69人妻影院| 在线观看人妻少妇| 免费看a级黄色片| 免费观看无遮挡的男女| 午夜免费观看性视频| 亚洲自拍偷在线| 亚洲国产最新在线播放| 人妻一区二区av| 国产精品av视频在线免费观看| 五月天丁香电影| 国产伦一二天堂av在线观看| 国产亚洲91精品色在线| 国产一级毛片在线| 两个人的视频大全免费| 亚洲av一区综合| 又爽又黄a免费视频| 国产永久视频网站| 国语对白做爰xxxⅹ性视频网站| 美女被艹到高潮喷水动态| 亚洲成人精品中文字幕电影| 国产淫片久久久久久久久| 久久久久免费精品人妻一区二区| 日韩三级伦理在线观看| 中文字幕免费在线视频6| 亚洲欧美一区二区三区黑人 | 日本免费在线观看一区| 久久99热这里只频精品6学生| 国产色爽女视频免费观看| freevideosex欧美| 国产亚洲av嫩草精品影院| 亚洲欧美成人精品一区二区| 免费av毛片视频| 亚洲精品第二区| 人人妻人人澡人人爽人人夜夜 | 亚洲人成网站高清观看| 少妇被粗大猛烈的视频| 亚洲精品第二区| www.av在线官网国产| 亚洲精品aⅴ在线观看| 91精品伊人久久大香线蕉| 欧美激情在线99| 国产精品嫩草影院av在线观看| 三级国产精品片| 久久精品久久久久久久性| 欧美成人精品欧美一级黄| 日本熟妇午夜| 久久国产乱子免费精品| 综合色丁香网| 一级av片app| 日韩国内少妇激情av| 深夜a级毛片| 日本爱情动作片www.在线观看| 国产精品伦人一区二区| 不卡视频在线观看欧美| 亚洲精品成人久久久久久| 免费观看无遮挡的男女| 亚洲av电影不卡..在线观看| 九草在线视频观看| 免费播放大片免费观看视频在线观看| 狂野欧美白嫩少妇大欣赏| 精品酒店卫生间| 欧美另类一区| 国产色婷婷99| 国产精品久久久久久精品电影| 一级黄片播放器| 国产老妇女一区| 一级二级三级毛片免费看| 国产 一区 欧美 日韩| 国产真实伦视频高清在线观看| 高清欧美精品videossex| 91狼人影院| 爱豆传媒免费全集在线观看| 精品99又大又爽又粗少妇毛片| 久久热精品热| 性色avwww在线观看| 又黄又爽又刺激的免费视频.| 欧美日韩综合久久久久久| 亚洲精品,欧美精品| 噜噜噜噜噜久久久久久91| 狂野欧美白嫩少妇大欣赏| 夫妻性生交免费视频一级片| 亚洲av电影在线观看一区二区三区 | 99热6这里只有精品| 精品国产露脸久久av麻豆 | 波多野结衣巨乳人妻| 免费av毛片视频| 麻豆av噜噜一区二区三区| 免费观看在线日韩| 97热精品久久久久久| 欧美+日韩+精品| 国产成人精品婷婷| 国产高清三级在线| 99热网站在线观看| 亚洲精品乱码久久久v下载方式| 亚洲精品亚洲一区二区| 国语对白做爰xxxⅹ性视频网站| 欧美人与善性xxx| 亚洲av福利一区| 午夜福利成人在线免费观看| 国产一区二区三区av在线| 在线免费十八禁| 欧美最新免费一区二区三区| 国产综合精华液| 国产欧美日韩精品一区二区| 国产男人的电影天堂91| 观看免费一级毛片| 亚洲综合色惰| 直男gayav资源| 精品人妻视频免费看| 国产精品一区www在线观看| 精品国产一区二区三区久久久樱花 | 亚洲经典国产精华液单| 国精品久久久久久国模美| a级一级毛片免费在线观看| 七月丁香在线播放| 99久久人妻综合| 国产 一区精品| 国产黄色视频一区二区在线观看| 纵有疾风起免费观看全集完整版 | 亚洲在久久综合| 亚洲熟妇中文字幕五十中出| 国产在视频线精品| 亚州av有码| 卡戴珊不雅视频在线播放| 欧美潮喷喷水| 亚洲欧美成人综合另类久久久| 99久久精品国产国产毛片| 黄色欧美视频在线观看| 欧美日韩精品成人综合77777| 日本欧美国产在线视频| 中文字幕av在线有码专区| 日本色播在线视频| 卡戴珊不雅视频在线播放| 亚洲图色成人| 寂寞人妻少妇视频99o| 丝袜美腿在线中文| 日韩,欧美,国产一区二区三区| 噜噜噜噜噜久久久久久91| 欧美97在线视频| 久久久久久久久大av| 久久久久久久大尺度免费视频| 国产亚洲5aaaaa淫片| 国产人妻一区二区三区在| 欧美区成人在线视频| 亚洲aⅴ乱码一区二区在线播放| 亚洲无线观看免费| 欧美性感艳星| 国产伦在线观看视频一区| 色综合站精品国产| 午夜视频国产福利| 一级毛片电影观看| 少妇的逼水好多| 国产亚洲91精品色在线| 亚洲在线自拍视频| 爱豆传媒免费全集在线观看| 国产精品麻豆人妻色哟哟久久 | 99久国产av精品| 免费看a级黄色片| 狂野欧美白嫩少妇大欣赏| av在线蜜桃| 春色校园在线视频观看| 成年女人看的毛片在线观看| 日韩欧美精品免费久久| 精品国产露脸久久av麻豆 | 久久久久久久久久成人| 青青草视频在线视频观看| 97超碰精品成人国产| 国产一区二区亚洲精品在线观看| 波多野结衣巨乳人妻| 精品久久久久久久久av| 亚洲av中文字字幕乱码综合| 丰满乱子伦码专区| 在线播放无遮挡| 男女国产视频网站| 久久久久久久久中文| 有码 亚洲区| 联通29元200g的流量卡| 国产黄a三级三级三级人| 国产成人精品一,二区| 亚洲av.av天堂| 亚洲图色成人| 色吧在线观看| 亚洲在线观看片| 在线免费观看的www视频| 尤物成人国产欧美一区二区三区| 最近最新中文字幕免费大全7| 别揉我奶头 嗯啊视频| 国内揄拍国产精品人妻在线| 亚洲久久久久久中文字幕| 3wmmmm亚洲av在线观看| 精品久久久久久成人av| 日韩伦理黄色片| 中文字幕制服av| 亚洲欧美精品专区久久| 欧美日韩国产mv在线观看视频 | 免费看av在线观看网站| 不卡视频在线观看欧美| 亚洲18禁久久av| 日本黄色片子视频| 特级一级黄色大片| 日韩精品有码人妻一区| 非洲黑人性xxxx精品又粗又长| 国产永久视频网站| 一级黄片播放器| 亚洲精品久久久久久婷婷小说| 日韩 亚洲 欧美在线| 在线观看av片永久免费下载| 麻豆成人午夜福利视频| 成人毛片60女人毛片免费| 好男人在线观看高清免费视频| 亚洲av日韩在线播放| 在线 av 中文字幕| 边亲边吃奶的免费视频| 国产中年淑女户外野战色| 亚洲av中文av极速乱| 97精品久久久久久久久久精品| 亚洲激情五月婷婷啪啪| 亚洲精品第二区| 亚洲精品久久久久久婷婷小说| 人妻系列 视频| 亚洲真实伦在线观看| 久久精品久久久久久久性| 91aial.com中文字幕在线观看| 干丝袜人妻中文字幕| 91aial.com中文字幕在线观看| 国产精品伦人一区二区| 日韩不卡一区二区三区视频在线| 亚洲天堂国产精品一区在线| 最近最新中文字幕免费大全7| 免费黄频网站在线观看国产| 日韩av免费高清视频| 日韩欧美一区视频在线观看 | 久久精品国产亚洲av涩爱| 国产黄色小视频在线观看| 伊人久久精品亚洲午夜| 毛片一级片免费看久久久久| 欧美日韩一区二区视频在线观看视频在线 | 亚洲av电影在线观看一区二区三区 | 成年版毛片免费区| 久久久久九九精品影院| 男的添女的下面高潮视频| 国产黄色视频一区二区在线观看| 国产黄频视频在线观看| 狠狠精品人妻久久久久久综合| 草草在线视频免费看| 乱码一卡2卡4卡精品| 美女黄网站色视频| 久久这里有精品视频免费| 国产一区二区亚洲精品在线观看| 国产伦在线观看视频一区| 久久精品久久精品一区二区三区| 夫妻性生交免费视频一级片| 国产精品嫩草影院av在线观看| 国产精品综合久久久久久久免费| 国产亚洲一区二区精品| 在线观看一区二区三区| 天堂影院成人在线观看| 国产亚洲91精品色在线| 久久亚洲国产成人精品v| 69人妻影院| 亚洲经典国产精华液单| 校园人妻丝袜中文字幕| 看黄色毛片网站| 麻豆国产97在线/欧美| 青春草视频在线免费观看| 亚洲综合色惰| 国产91av在线免费观看| 日韩欧美精品v在线| 国产精品一二三区在线看| 深夜a级毛片| 欧美成人午夜免费资源| av又黄又爽大尺度在线免费看| 日韩欧美一区视频在线观看 | 国产男女超爽视频在线观看| 成年人午夜在线观看视频 | 精品久久久精品久久久| 欧美精品一区二区大全| 亚洲一区高清亚洲精品| 国产不卡一卡二| 日韩欧美国产在线观看| 男女边吃奶边做爰视频| 日韩国内少妇激情av| 精品久久久久久久末码| 麻豆久久精品国产亚洲av| 深夜a级毛片| 熟女人妻精品中文字幕| 中文欧美无线码| 蜜桃久久精品国产亚洲av| 日韩大片免费观看网站| 亚洲av中文字字幕乱码综合| 一夜夜www| 久久久久久久久久人人人人人人| 国产老妇伦熟女老妇高清| 久久久午夜欧美精品| 91精品国产九色| 免费av毛片视频| 联通29元200g的流量卡| 一级毛片久久久久久久久女| 人人妻人人澡人人爽人人夜夜 | 少妇高潮的动态图| 美女cb高潮喷水在线观看| 日韩人妻高清精品专区| 亚洲精品456在线播放app| 久久久成人免费电影| 免费观看精品视频网站| 少妇裸体淫交视频免费看高清| 亚洲av二区三区四区| 亚洲综合精品二区| 在线观看av片永久免费下载| 欧美成人精品欧美一级黄| av免费在线看不卡| 久久久久久久久中文| 两个人视频免费观看高清| 高清午夜精品一区二区三区| 久久精品夜色国产| 男女下面进入的视频免费午夜| 老师上课跳d突然被开到最大视频| 国产精品三级大全| 天天一区二区日本电影三级| 亚洲精品国产av成人精品| 91午夜精品亚洲一区二区三区| 国产v大片淫在线免费观看| 一级毛片 在线播放| 亚洲av成人精品一区久久| 淫秽高清视频在线观看| 国产黄片美女视频| 成人性生交大片免费视频hd| 亚洲精品日韩av片在线观看| 国产一区有黄有色的免费视频 | 午夜精品一区二区三区免费看| 嫩草影院精品99| 亚洲av二区三区四区| 在线免费十八禁| 少妇人妻一区二区三区视频| 国产高清三级在线| 深爱激情五月婷婷| 国产有黄有色有爽视频| 少妇熟女欧美另类| 色综合亚洲欧美另类图片| av专区在线播放| av在线播放精品| 国产免费福利视频在线观看| 亚洲一区高清亚洲精品| 我的女老师完整版在线观看| 亚洲国产日韩欧美精品在线观看| 97热精品久久久久久| 久久精品人妻少妇| 亚洲不卡免费看| 少妇的逼好多水| 国产一区二区三区综合在线观看 | 国模一区二区三区四区视频| 日韩精品青青久久久久久| 熟妇人妻不卡中文字幕| 日韩av免费高清视频| 国产成人freesex在线| 大香蕉97超碰在线| 少妇人妻一区二区三区视频| 亚洲在线观看片| 综合色丁香网| 国产一级毛片七仙女欲春2| 欧美3d第一页| 亚州av有码| 高清视频免费观看一区二区 | 只有这里有精品99| 国产综合懂色| 欧美日韩视频高清一区二区三区二| 熟女电影av网| 干丝袜人妻中文字幕| 97超视频在线观看视频| 伦精品一区二区三区| 精品久久久精品久久久| 亚洲精品,欧美精品| 在线观看人妻少妇| 一级毛片久久久久久久久女| 午夜福利在线观看免费完整高清在| 韩国高清视频一区二区三区| 久久久久久久久大av| 亚洲国产日韩欧美精品在线观看| 亚洲综合精品二区| 亚洲成人精品中文字幕电影| 天堂网av新在线| 一夜夜www| 欧美区成人在线视频| 99久久精品国产国产毛片| 成人毛片a级毛片在线播放| 精品人妻熟女av久视频| 国产亚洲av片在线观看秒播厂 | av卡一久久| 在线a可以看的网站| 精品久久久精品久久久| 精品酒店卫生间| 一区二区三区免费毛片| 在线天堂最新版资源| 成人鲁丝片一二三区免费| 国产成年人精品一区二区| 不卡视频在线观看欧美| 婷婷色综合www| 国产一区二区三区av在线| 成人美女网站在线观看视频| 国产黄色小视频在线观看| 亚洲欧美一区二区三区黑人 | 免费不卡的大黄色大毛片视频在线观看 | 亚洲自偷自拍三级| av在线亚洲专区| 午夜福利视频1000在线观看| av免费观看日本| 成人欧美大片| 欧美xxxx黑人xx丫x性爽| 久久99精品国语久久久| 真实男女啪啪啪动态图| 日本爱情动作片www.在线观看| 91久久精品电影网| 久久久久久久久久久丰满| 欧美性感艳星| 我的老师免费观看完整版| 亚洲av不卡在线观看| 色综合站精品国产| 精品国产露脸久久av麻豆 | 插逼视频在线观看| 人妻少妇偷人精品九色| 国产乱人偷精品视频| 五月伊人婷婷丁香| 免费黄网站久久成人精品| 七月丁香在线播放| 国产在线男女| 一二三四中文在线观看免费高清| 三级国产精品欧美在线观看| 亚洲成人一二三区av| 九九久久精品国产亚洲av麻豆| 天堂影院成人在线观看| 久久久精品94久久精品| 成年版毛片免费区| 简卡轻食公司| 美女主播在线视频| 欧美一级a爱片免费观看看| 少妇裸体淫交视频免费看高清| 在线天堂最新版资源| 99热这里只有是精品在线观看| 日韩制服骚丝袜av| 亚洲在线观看片| xxx大片免费视频| 日韩一区二区视频免费看| a级毛片免费高清观看在线播放| 国产成人午夜福利电影在线观看| 一夜夜www| 亚洲成人av在线免费| 能在线免费观看的黄片| 国产精品无大码| 久久精品国产鲁丝片午夜精品| 2018国产大陆天天弄谢| 国产成人一区二区在线| 亚洲婷婷狠狠爱综合网| 一区二区三区高清视频在线| 成人二区视频| 色吧在线观看|