• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    YOLOv2PD:An Efficient Pedestrian Detection Algorithm Using Improved YOLOv2 Model

    2021-12-15 07:07:00ChintakindiBalaramMurthyMohammadFarukhHashmiGhulamMuhammadandSalmanAlQahtani
    Computers Materials&Continua 2021年12期

    Chintakindi Balaram Murthy,Mohammad Farukh Hashmi,Ghulam Muhammadand Salman A.AlQahtani

    1Department of Electronics and Communication Engineering,National Institute of Technology,Warangal,506004,India

    2Research Chair of Pervasive and Mobile Computing,King Saud University,Riyadh,11543,Saudi Arabia

    3Department of Computer Engineering,College of Computer and Information Sciences,King Saud University,Riyadh,11543,Saudi Arabia

    Abstract:Real-time pedestrian detection is an important task for unmanned driving systems and video surveillance.The existing pedestrian detection methods often work at low speed and also fail to detect smaller and densely distributed pedestrians by losing some of their detection accuracy in such cases.Therefore, the proposed algorithm YOLOv2 (“YOU ONLY LOOK ONCE Version 2”)-based pedestrian detection (referred to as YOLOv2PD) would be more suitable for detecting smaller and densely distributed pedestrians in real-time complex road scenes.The proposed YOLOv2PD algorithm adopts a Multi-layer Feature Fusion (MLFF) strategy, which helps to improve the model’s feature extraction ability.In addition,one repeated convolution layer is removed from the final layer,which in turn reduces the computational complexity without losing any detection accuracy.The proposed algorithm applies the K-means clustering method on the Pascal Voc-2007+2012 pedestrian dataset before training to find the optimal anchor boxes.Both the proposed network structure and the loss function are improved to make the model more accurate and faster while detecting smaller pedestrians.Experimental results show that,at 544×544 image resolution,the proposed model achieves 80.7%average precision(AP),which is 2.1%higher than the YOLOv2 Model on the Pascal Voc-2007+2012 pedestrian dataset.Besides,based on the experimental results, the proposed model YOLOv2PD achieves a good trade-off balance between detection accuracy and real-time speed when evaluated on INRIA and Caltech test pedestrian datasets and achieves state-of-the-art detection results.

    Keywords:Computer vision; K-means clustering; multi-layer feature fusion strategy; pedestrian detection; YOLOv2PD

    Abbreviations

    AP Average Precision

    CV Computer Vision

    CUDA Compute Unified Device Architecture

    DPM Deformable Part Model

    FPS Frames per second

    FP False Positive

    FN False Negative

    HOG Histogram of Oriented Gradient

    IoU Intersection over Union

    MR Miss-rate

    MLFF Multi-layer Feature Fusion

    Pascal Voc Pascal Visual Object Classes

    RCNN Regions Based Convolutional Neural Networks

    SPPNet Spatial Pyramid Pooling Network

    SSD Single Shot Multi-Box Detector

    SOTA State-of-the-art

    TP True Positive

    TN True Negative

    YOLO YOU ONLY LOOK ONCE

    YOLOv2 YOU ONLY LOOK ONCE Version 2

    YOLOv2PD YOU ONLY LOOK ONCE Version 2 Based Pedestrian Detection

    1 Introduction

    One of the most important applications of Computer vision (CV) in self-driving cars is pedestrian detection.The field of pedestrian detection covers video surveillance, criminal investigations, self-driving cars, and robotics.Real-time pedestrian detection is an important task for unmanned driving systems.The vision system of autonomous vehicle technology was initially very difficult to develop in the field of CV; however, owing to continuous improvements of hardware computational power, many researchers have attempted to develop reliable vision systems for selfdriving cars.Since 2012, deep learning has been developed and achieved tremendous progress in the field of CV.In the field of artificial intelligence, many deep learning-based algorithms have been introduced and used in a wide range of applications, such as in signal, audio, image, and video processing.In particular, deep learning-based algorithms play a groundbreaking role in fields such as image and video processing, for example, image classification and detection.

    One of the direct applications of real-time pedestrian detection is that it should automatically locate pedestrians accurately with on-shelf cameras, since it plays a crucial role in robotics and unmanned driving systems.Despite tremendous progress having been achieved recently, this task still remains challenging due to the complexity of road scenes, such as them being crowded,occluded, containing deformations and exhibiting lighting changes.Currently, unmanned driving systems are among the major fields of research in CV, for which the real-time detection of pedestrians is essential to avoid possible accidents.Although deep learning-based techniques improve detection accuracy, there is still a huge gap between human and machine perception [1].A complex background, low-resolution images, lighting conditions, and occluded and distant smaller objects reduces the model accuracy.To date, most researchers in this field have focused only on color-image-based object detection.Therefore, when detecting objects in a shadowy environment or objects captured at night, lower detection accuracy is achieved.

    This is the major drawback of reliable vision-based detection systems since self-driving cars in real-time extremely complex environments should be able to detect objects in the daytime or at night.Nevertheless, current state-of-the-art (SOTA) real-time pedestrian detection still falls short of the fast and accurate human perception levels [2].

    Currently, pedestrian detection methods are classified into two time slots:traditional and deep learning time slot methods.Traditional time slot methods cover various traditional machine learning algorithms such as Voila Jones detector [3], Deformable part model (DPM) [4], Histogram of oriented gradient (HOG) [5] and multi-scale gradient histograms [6].These methods are timeconsuming, require complex steps, are expensive, and require a high level of human interference.In the recent evolution of deep learning techniques since 2012, such techniques have become very popular and deep CNN-based pedestrian detection methods have achieved better performance than traditional time slot methods [7,8].The first deep learning-based object detection model was RCNN [9].This method generates a region of interest by using a selective search window for deep learning-based object detection, as implemented in all RCNN series.Deep learning time slot methods cover both two-stage detectors such as RCNN [9], SPPNet [10], Fast-RCNN [11], Faster RCNN [12] and Mask-RCNN [13] and single-stage detectors such as SSD [14] and YOLO [15].Therefore, in the current scenario for real-time pedestrian detection, these methods are not quite suitable.

    Generally, the speed of deep learning-based object detection methods is low, with these methods being unable to meet real-time requirements of self-driving cars.Therefore, to improve both speed and detection accuracy, Redmon et al.[15] proposed the YOLO network, a single endto-end object regression framework.Later, Redmon et al.[16] implemented YOLOv2 to overcome the drawbacks of the YOLO [15] framework.YOLOv2 [16] improves the speed of the detection algorithm without losing any part of the detection accuracy.However, when detecting smaller objects in complex environments, it achieves low detection accuracy.

    To improve both detection accuracy and speed when detecting smaller and densely distributed pedestrians, a new pedestrian detection technique is proposed, YOLOv2-based pedestrian detection(in short, YOLOv2PD).An efficient K-means clustering [17] algorithm is applied to select six different anchor box sizes while training the Pascal Voc-2007+2012 pedestrian dataset.

    The contributions of the proposed work can be summarized as follows:

    (1) The proposed YOLOv2PD model adopts the MLFF strategy to improve the model’s feature extraction ability and, at the higher end, one convolution layer is eliminated.

    (2) Moreover, intuitively, to test the effectiveness of the proposed model, another model referred to as YOLOv2 Model A is implemented and compared.

    (3) The loss function is improved by applying normalization, which reduces the effect of different pedestrian sizes in an image, and which potentially optimizes the detected bounding boxes.

    (4) Through qualitative and quantitative experiments conducted on Pascal Voc-2007+2012 Pedestrian, INRIA and Caltech pedestrian datasets, we validate the effectiveness of our algorithm, showing that it has better detection performance on smaller pedestrians.

    The rest of the paper is organized as follows.Sections 2 covers related work.In Section 3,the proposed YOLOv2PD algorithm is illustrated.Section 4 covers the benchmark datasets Pascal Voc-2007+2012 Pedestrian, INRIA and Caltech; the experimental results and analysis are discussed.Finally, the conclusion is presented and future works are discussed.

    2 Related Work

    The research field of pedestrian detection has existed for several decades, in which different technologies have been employed for this detection, many of which have had significant impacts.Some methods aim to improve the basic features utilized [18–20], while others are intended to optimize the detection algorithms [21,22], while some other methods incorporate DPM [23] or use the advantage of context [23,24].

    Benenson et al.[18] evaluated the complete performance of multifarious features and methods.Benenson et al.[20] implemented the fastest technique to achieve a frame rate of 100 frames per second (FPS) for pedestrian detection.After 2012, the deep learning era started, which has greatly improved the accuracy of pedestrian detection [21,24–26].However, their run time on each image is slightly or markedly slower, taking a few seconds.Moreover, many remarkable techniques are now employed in CNNs.Paisitkriangkrai et al.[25] proposed new features constructed based on low-level vision features and incorporated spatial pooling to improve translational invariance which in turn improves the robustness of pedestrian detection process.The ConvNet [27] method uses convents for detecting pedestrians.It employs convolutional sparse coding to initialize each layer at the start and later performs fine-tuning to perform object detection.RPN-BF [28] is a perfect fusion of Region Proposal Networks (RPN) and Boosted Forest Classifier.RPN proposed in Faster RCNN [12] generates candidate bounding boxes, high-resolution feature maps, and confidence scores.To shape the Boosted Forest Classifier, it also employs the Real-boost algorithm for using the obtained information from RPN.This two-stage detector has shown good performance results on pedestrian test datasets.Murthy et al.[29] presented a study of pedestrian detection using various custom-made deep learning techniques.

    Li et al.[30] proposed a network structure which integrates both region generation and prediction modules for accurate localization of real-time small-scale pedestrian detection.Li et al.[31]proposed scale-aware Fast-RCNN method for detecting pedestrians of various scales, and applied anchor box mechanism onto multiple feature layers.In addition, Ouyang et al.[32] proposed a unified deep neural network for jointly learning four key components, namely, feature extraction+ deformation + occlusion and classification for pedestrian detection.Pang et al.[33] introduced a mask-guided attention network for detecting occluded pedestrians, which emphasizes only visible regions and suppresses occluded regions by modulating full body features.However, this method fails to achieve satisfactory results on heavily occluded pedestrians.Zhang et al.[34] proposed a simple and compact method by incorporating a channel-wise attention network on Faster RCNN detector while detecting occluded pedestrians.

    Song et al.[35] proposed a novel method by integrating both somatic topological line localization and temporal feature aggregation for detecting small-scale pedestrians, which are relatively far from the camera.This method also eliminates ambiguities in occluded pedestrians by introducing a post-processing scheme based on Markov Random Field (MRF).Zhang et al.[36] proposed “keypoint-guided super-resolution network” (KGSNet) for detecting small-scale and heavily occluded pedestrians.Initially, this network is trained to generate a super-resolution pedestrian image and then a part estimation module encodes the semantic information of four human body parts.

    Lin et al.[37] proposed a graininess-aware feature learning method for detecting small-scale and occluded pedestrians.Attention mechanism is used to generate graininess-aware feature maps and then to enhance the features, a zoom-in-zoom-out module is introduced.Wu et al.[38]proposed a novel self-mimic loss learning method, to improve the detection accuracy of smallscale pedestrians.Hsu et al.[39] proposed a new ratio-and-scale-aware YOLO (RSA-YOLO)and achieves extremely better results while detecting small-pedestrians.Moreover, Han et al.[40]proposed a novel small-scale sense (SSN) network, which can generate some proposal regions and is effective when detecting small-scale pedestrians.

    Specifically, two-stage deep learning-based object detectors offer advantages in achieving both higher localization accuracy and precision.The process requires huge resources and yet the computational efficiency is low.Owing to the unified network structures, one-stage detectors are much faster than two-stage detectors, even though the model precision decreases.Moreover, the amount of training data plays a vital role in deep learning-based object detectors.We present an end-to-end single deep neural network for detecting smaller and densely distributed pedestrians in real time inspired by YOLOv2.YOLOv2 (“You only look once version 2”) [16] is an endto-end single deep neural network that integrates feature extraction, bounding box extraction,object classification and detection.YOLOv2 is adopted as a basic model in order to achieve accuracy and higher speed when detecting smaller and densely distributed pedestrians.After making modifications in the YOLOv2 network structure and hyperparameters, it was adopted for the accurate detection of smaller and densely distributed pedestrians.

    The proposed method YOLOv2PD adopts the YOLOv2 deep learning framework [16] as a base model and hyperparameters are adjusted to achieve better detection accuracy in real time.Additionally, at the higher end, some unwanted repeated convolution layers are eliminated in the proposed model, so it consumes less computational time than the YOLOv2 Model.Therefore, the YOLOv2PD model is the best method for accurate real-time detection of smaller and densely distributed pedestrians.The proposed model performance is evaluated on the Pascal Voc-2007+2012 Pedestrian dataset and its performance is compared with YOLOv2 and YOLOv2 Model A models.To test the robustness of the proposed model, YOLOv2PD is also evaluated on both INRIA [5] and Caltech [41] pedestrian datasets.

    3 YOLOv2PD Proposed Algorithm

    3.1 Anchor Boxes Selected Based on K-means Clustering

    The proposed method applies a K-means clustering algorithm on the Pascal Voc-2007+2012 pedestrian dataset during training and selects the optimal number of anchor boxes of different sizes.It works by replacing traditional Euclidean distance with the distance function of YOLOv2 while implementing the K-means clustering algorithm.Therefore, the error obtained is made irrelevant with respect to anchor box sizes by adopting IoU as an evaluation metric, as shown in Eq.(1).

    where box is the sample; centroid is cluster center point; IoU (box, centroid) is the overlap ratio between cluster and center boxes.Based on the clustering results analysis, the K value was chosen to be 6; therefore, six different anchor box sizes would be applied in order to improve the positioning accuracy.Finally, by implementing the K-means clustering algorithm on the training dataset, a suitable number of different anchor box sizes are selected for pedestrian detection,which in turn improves the positioning accuracy.

    3.2 Improved Loss Function

    Since images are captured using a video surveillance camera, some of the pedestrian images might be bigger, with pedestrians being nearer the camera, while some pedestrian images might be smaller, with pedestrians being located far away from the camera during detection.Therefore,pedestrians would appear smaller in the image when they are far from the camera, and vice versa.As such, sizes may vary in the captured images, even though the pedestrian is identical.

    During YOLOv2 training, objects of different sizes show different effects on the network and produce large errors, particularly for images with smaller and densely distributed objects.To overcome this drawback, loss calculation for bounding box (BB) width and height is improved by applying normalization.Eq.(2) shows the improved loss function as:

    3.3 Network Design

    Multi-layer Feature Fusion (MLFF) Approach:In pedestrian detection, variations among pedestrians include occlusion, illumination changes, color, height, and contour, whereas local features exist only in the lower layers of CNN.Therefore, to use local features fully, an MLFF approach was implemented in YOLOv2PD.The Reorg aim is to keep feature maps of those layers the same.Part (a) passes through the following 3×3 and 1×1 convolution layers and then a down-sampling factor of Reorg/8 is applied, as shown in Fig.1.Similarly, part (b) and part (c)perform the same operations, but with down-sampling factors of 4 and 2, respectively.Part (a),(b) local features, and part (c) global features of one layer are fused.This is done so that the network would distinguish the tiny differences among pedestrians and also it improves the network understanding of local features.

    YOLOv2 is a fast and accurate object detection model.The YOLOv2 network can detect 9000 classes and variations among multiple objects are wider, such as cell phones, cars, fruits,sofas, and dogs.There are three repeated 3 × 3 × 1024 convolutional layers in the YOLOv2 network.Generally, at the higher end, repeated convolution operation deals with multiple classes and widely differing objects, such as fruits, animals, and vehicles.However, our main concern is only detecting the pedestrian class and feature differences among pedestrians are minute.Thus,the model performance may not improve due to repeated convolution layers at the higher end and, due to their presence, the model becomes more complex.Therefore, repeated convolution layers are removed from the higher end in the proposed models.This strategy would achieve almost competitive performance and reduce the time complexity of the Yolov2 network.Thus,three repeated 3×3×1024 convolution layers are reduced to two in the proposed model, as shown in Fig.1.

    Figure 1:YOLOv2PD network architecture

    A novel YOLOv2PD network structure is designed by adopting the MLFF approach and one unwanted convolutional layer is removed at the higher end.Moreover, intuitively, to test the effectiveness of the proposed model, another model, referred to as YOLOv2 Model A, was implemented and compared.The YOLOv2 Model A removed two 3×3×1024 convolution layers and the YOLOv2PD model removed only one 3×3×1024 convolution layer when compared with the YOLOv2 network.Tab.1 shows the comparison between YOLOv2, YOLOv2 Model A, and YOLOv2PD network architecture.

    Table 1:YOLOv2, YOLOv2 Model A, and YOLOv2PD network architecture

    4 Datasets and Experimental Results

    4.1 Datasets

    Pascal Voc-2007+2012 dataset [42]:This dataset contains 20 object classes and around 17,125 labeled images; it is a complete dataset generally used for object detection and classification.An unsupervised learning method (K-means clustering) is applied during training.Since manual annotation of a dataset is a complex and huge project, around 10,080 pedestrian and nonpedestrian images (referred to as the Pascal Voc-2007+2012 Pedestrian dataset) were extracted from Pascal dataset [42].

    The INRIA Pedestrian dataset [5] contains 1826 pedestrians, with image resolution 64×128.The pedestrian images captured in this dataset possess a complex background, illumination changes, various degrees of occlusion, variations in human posture, and individuals wearing different clothes.

    The Caltech pedestrian dataset [41] contains a set of video sequences of 640×480 in size captured from an urban environment.It includes training (set 00 to set 05) subsets and testing (set 06 to set 10) subsets.It contains 250 k video frames, 350 k bounding boxes and 2.3 k pedestrians(“person” or “people” labels) are annotated.The training dataset is formed by extracting every image after every 30 frames from set 00 to set 05 and testing images are extracted from set 06 to set 10.Tab.2 shows the datasets used for both training and testing of the proposed algorithm.

    Table 2:Datasets used by the proposed algorithm for Training & Testing

    4.2 Experimental Setup

    The experiments were carried out on a workstation during the training phase; the testing phase was also performed on the same workstation.Darknet was chosen as a feature extractor for all of the models, which was trained on a huge ImageNet dataset.The experimental setup of the workstation is Windows 10 pro OS, Intel Xeon 64-bit CPU @3.60 GHz, 64 GB RAM, Nvidia Quadro P4000 GPU, CUDA 10.0 & CUDNN 7.4 GPU acceleration library and Tensorflow 1.x deep learning framework.

    4.3 Training and Evaluation Metrics

    The model training was carried out on Pascal Voc-2007+2012 Pedestrian dataset (9072)training images and tested on 1008 testing images, since we are only concerned with pedestrian images.The input image size is resized to 416×416 resolution and various data augmentation techniques are applied, such as color shifting, flipping, cropping, and random sampling, in order to enhance the training process.All of the three models are trained for 40 epochs, with an initial learning rate of 0.001, and later learning rate is divided by 10 at 60 and 80 epochs respectively.During the model training, it randomly selects a new input image of different resolution after every 20 epochs.Since multi-scale training strategy improves model robustness, so it can perform better prediction on images with different resolutions.While training, Caltech dataset, the original images are up-sampled to 1024×1024 pixels, one mini-batch contains 16 images, learning rate is 10-4 and the model training is stopped after 80 epochs.

    Average precision (AP) and inference speed (FPS-Frames per second) are the standard techniques preferred to evaluate the model performance.Intersection over union (IoU) is a good evaluation metric used to measure the accuracy of the designed model on a test dataset.IoU is simply computed as the area of intersection divided by the area of union.IoU helps to determine whether a predicted BB is a True Positive (TP), False positive (FP) or False Negative (FN) by defining a threshold of ≥0.5.

    Recall:A measure of how good the model is at finding all of the positives.Precision:A measure of the accuracy of our predictions.These two terms are inversely proportional to each other.

    AP:This is the area under the precision–recall curve, which shows the correlation between precision and recall at different confidence scores.A higher AP value indicates better detection accuracy.

    The performance of the model while validating INRIA and Caltech test datasets was visualized using a plot between the number of false positives per image and the miss rate (MR).The ratio between the number of FNs and the total number of positive samples (N) is referred to as the MR.

    There is another relationship between the miss rate and recall expressed as:

    4.4 Results and Analysis

    Fig.2 shows the analysis of the training stage of all three models.The y-axis indicates average loss and the x-axis indicates the number of iterations performed in training.It is clear from Fig.2 that the average loss curve is not stable up to approximately 10000 iterations.Compared with all of the other models, the average loss curve of the YOLOv2PD model decreases faster initially,followed by that of YOLOv2 Model A.The reason for this is that both YOLOv2PD and YOLOv2 Model A adopted a multi-layered feature fusion strategy, so they obtained more local features,which accelerated the training convergence.During the training stage, initially the YOLOv2PD model first reached a minimum average loss value (overall lowest value = 0.54), followed by YOLOv2 Model A and YOLOv2 models.Therefore, the YOLOv2PD model is more suitable for detecting small pedestrians on the Pascal Voc-2007+2012 pedestrian dataset.

    Figure 2:Analysis of training stage of all of the models

    Fig.3 shows the precisionvs.recall (PR) curve obtained on the Pascal Voc-2007+ 2012 pedestrian dataset of all three models.The graph shows that, with increasing recall value at the convergence point, the precision gradually starts decreasing.

    Figure 3:PR curves of all of the models on the Pascal Voc-2007+2012 pedestrian dataset

    With different input image resolutions of 416×416, 544×544, and 608×608, YOLOv2PD achieves comparable detection performance when compared with YOLOv2 Model A and YOLOv2.Tab.3 compares the detection performance of all models for different image resolutions with respect to AP and inference speed (FPS) parameters.The proposed network YOLOv2PD achieves AP, that is, detection performance of 79.5, 80.7, and 82.3 respectively.From these results,it is clear that, as the applied input image resolution increases, the AP value increases but at the same time inference speed decreases.

    To have a model that runs at higher inference speed, an image size of 416×416 is the best choice.As the input image size increases, inference speed decreases since these terms are directly proportional to each other.However, we are concerned with detecting smaller and densely distributed pedestrians, so 416×416 images are not quite suitable as they miss the detection of many smaller objects.Therefore, we consider selecting a 544×544 image size for detecting smaller and densely distributed pedestrians.From the experimental results, our proposed algorithm runs at 36.3 FPS in real time on 544×544 image resolution.In this study, if the AP is considered,then an image size of 544×544 would be the best choice as the proposed model achieves 80.7%detection accuracy, which is 2.1% higher than that of YOLOv2 [16].The proposed model runs at 30.6 FPS for the 608×608 image resolution, but the inference speed falls by 5.7 FPS compared to 544×544 image resolution.

    Table 3:Evaluation results of all of the models on the pedestrian test dataset (IoU@0.5)

    4.5 Small Pedestrian Detection

    The Pascal Voc-2007+2012 pedestrian dataset contains 20 different classes and every class may have small objects.We were concerned with detecting smaller and densely distributed pedestrians in this dataset, so we manually picked up 330 images that mainly included smaller pedestrians to evaluate the model performance.Fig.4 shows detection results of all models and compared with YOLOv3 [43] SOTA detector.From these detection results, it is evident that the proposed model can produce better prediction on smaller and densely distributed pedestrians than the other models.

    Figure 4:Detection results of YOLOv2, YOLOv2 Model A, YOLOv2PD and YOLOv3 Models

    The evaluation results of all three models on the INRIA test dataset are expressed in terms of average precision and inference speed (milliseconds).Tab.4 shows detected results on the INRIA test dataset for different image resolutions.At 544 ×544 test image resolution, the proposed model achieves 91.2% AP, which constitutes an improvement by 6.6% and 11.4% compared with YOLOv2 Model A and YOLOv2 models, respectively.This is because our model uses the MLFF strategy while detecting smaller pedestrians.

    Table 4:Detection results of all of the models on the INRIA Test dataset.(IoU@0.5)

    To test the robustness of the proposed model, we compared our model performance on the INRIA pedestrian test dataset with several SOTA algorithms.

    Tab.5 shows a comparison of the YOLOv2PD model performance with the advanced existing algorithms evaluated in terms of average MR and runtime (FPS) on a reasonable test dataset.Our model achieves better detection performance than YOLOv2 [16], Spatial Pooling [25] and Y-PD [44] and is improved by 4.7%, 3.4% and 1.3% respectively, but lags behind YOLOv3 [43] and F-DNN [45] by 0.6% and 1% respectively.Obviously, on the INRIA pedestrian test dataset, the proposed model achieves a better trade-off balance between speed and accuracy when detecting pedestrians.

    Table 5:Comparison of YOLOv2PD results with recent SOTA methods on the INRIA test dataset

    Tab.6 shows a comparison of the proposed model performance with the advanced existing algorithms on the Caltech test dataset, evaluated in terms of MR, average precision, and detection speed.

    Table 6:Comparison of YOLOv2PD detection results with recent SOTA methods on the Caltech test dataset (IoU@0.75)

    From Tab.6, it is clear that, on the Caltech test dataset, the proposed model has better detection performance than RPN + BF [28], SA-FastRCNN [31], UDN + SS [32], Faster RCNN+ ATT-Vbb [34], SSNet [40], Y-PD [44] and CompactACT + Deep [47], and models on the reasonable subset [h ∈(50, ∞)].However, the proposed model average miss rate falls behind those of M-GAN [33], TTL (MRF) + LSTM [35] and SDS-RCNN [46] models by 0.65%, 0.80% and 0.12% respectively.

    To show the findings more intuitively, regarding the real-time performance of the proposed algorithm to achieve a perfect balance between detection speed and accuracy, we fed a real-time test video to all models.The detection results of the randomly selected 79thframe for all of the models are shown in Fig.5.We evaluated the running time for these three models on a realtime input test video.The detection speed on an input image of size 544×544 was 32 FPS for YOLOv2, 38.2 FPS for YOLOv2 Model A, 36.3 FPS for YOLOv2PD and 20 FPS for YOLOv3.Although the proposed model runs in real-time, it fails to detect smaller and similar occluded pedestrians.The use of the Internet of Things may make the method more efficient [48].

    Figure 5:Real-time detection results of YOLOv2, YOLOv2 Model A, YOLOv2PD and YOLOv3 Models

    5 Conclusion

    A new advanced model named YOLOv2PD was proposed for the accurate detection of smaller and densely distributed pedestrians.The proposed network YOLOv2PD structure was designed to improve the network’s feature extraction ability by adopting the MLFF strategy and, at the higher end, one repeated convolutional layer was removed.To improve the detection accuracy while detecting smaller and more densely distributed pedestrians, the loss function was improved by applying normalization.The experimental results show that, for an applied input image of 544×544 in size, the proposed algorithm achieves 80.7% AP, which is 2.1% higher than that of the YOLOv2 Model on the Pascal Voc-2007+2012 pedestrian test dataset.To test the robustness of the proposed algorithm, we captured a real-time video and fed it images at 544×544 resolution; it obtained 36.3 FPS and achieved 80.7% detection accuracy compared with the SOTA YOLOv2 Model.The experimental results show that the proposed model achieves 7.8 average MR on INRIA and 0.381 AP on Caltech pedestrian test datasets.Although the model was run in real time, there is still room for improvement of the speed, miss rate on INRIA test dataset and miss detection of small similar and occluded pedestrians.

    Funding Statement:The authors are grateful to the Deanship of Scientific Research, King Saud University, Riyadh, Saudi Arabia, for funding this work through the Vice Deanship of Scientific Research Chairs:Research Chair of Pervasive and Mobile Computing.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    中国美白少妇内射xxxbb| 2022亚洲国产成人精品| 国产亚洲午夜精品一区二区久久| 好男人视频免费观看在线| 51国产日韩欧美| 十八禁高潮呻吟视频 | 亚洲成人手机| a级毛色黄片| 国产日韩欧美亚洲二区| .国产精品久久| av免费观看日本| 色哟哟·www| 国产亚洲精品久久久com| 色视频www国产| 秋霞伦理黄片| 男女免费视频国产| 亚洲内射少妇av| 一级毛片我不卡| 国产91av在线免费观看| 日韩熟女老妇一区二区性免费视频| 欧美亚洲 丝袜 人妻 在线| 国产精品欧美亚洲77777| 老司机影院成人| 中文在线观看免费www的网站| 岛国毛片在线播放| 人人妻人人澡人人看| 美女xxoo啪啪120秒动态图| 下体分泌物呈黄色| 亚洲丝袜综合中文字幕| 男女啪啪激烈高潮av片| av天堂久久9| 亚洲精品日本国产第一区| 涩涩av久久男人的天堂| 亚洲电影在线观看av| 国产精品秋霞免费鲁丝片| 天美传媒精品一区二区| 18禁动态无遮挡网站| 亚洲无线观看免费| 国产熟女午夜一区二区三区 | 亚洲成人一二三区av| 亚洲精品一二三| 久久久亚洲精品成人影院| 国产亚洲欧美精品永久| 国产 精品1| 久久久久久久亚洲中文字幕| 久久99一区二区三区| 国产欧美另类精品又又久久亚洲欧美| av女优亚洲男人天堂| 久久久久久久亚洲中文字幕| 久久热精品热| 欧美性感艳星| 蜜桃久久精品国产亚洲av| 少妇高潮的动态图| 日日撸夜夜添| 成年人午夜在线观看视频| 精品一区二区三卡| 成人毛片a级毛片在线播放| xxx大片免费视频| 国产精品久久久久成人av| 丁香六月天网| 大陆偷拍与自拍| www.色视频.com| 美女内射精品一级片tv| 国产色爽女视频免费观看| 欧美日韩一区二区视频在线观看视频在线| 中文精品一卡2卡3卡4更新| 亚洲四区av| 精品久久久噜噜| 亚洲欧美一区二区三区黑人 | 精品久久久精品久久久| 三上悠亚av全集在线观看 | 国产片特级美女逼逼视频| 国产免费又黄又爽又色| 亚洲精品亚洲一区二区| 欧美日韩视频高清一区二区三区二| 免费在线观看成人毛片| 久久久久久久久久成人| 国产一区二区在线观看日韩| 最近手机中文字幕大全| h视频一区二区三区| 91久久精品电影网| 欧美国产精品一级二级三级 | 国产欧美日韩综合在线一区二区 | 欧美+日韩+精品| 欧美日韩在线观看h| 天美传媒精品一区二区| 91成人精品电影| 一级毛片黄色毛片免费观看视频| 国产亚洲午夜精品一区二区久久| 嘟嘟电影网在线观看| 99re6热这里在线精品视频| 亚洲熟女精品中文字幕| 午夜福利影视在线免费观看| 国产成人午夜福利电影在线观看| 国产午夜精品久久久久久一区二区三区| 精品国产乱码久久久久久小说| 观看美女的网站| 欧美日韩亚洲高清精品| 你懂的网址亚洲精品在线观看| 少妇人妻久久综合中文| 欧美日韩av久久| 亚洲精品一二三| 国产 一区精品| 国产高清有码在线观看视频| 久久精品国产a三级三级三级| 在线天堂最新版资源| 国产一区亚洲一区在线观看| 黄色视频在线播放观看不卡| 亚洲在久久综合| 日韩中字成人| av不卡在线播放| 国产淫语在线视频| 日韩中字成人| 中文乱码字字幕精品一区二区三区| 夜夜爽夜夜爽视频| 国产午夜精品久久久久久一区二区三区| 久久97久久精品| 99久久人妻综合| 中文字幕人妻丝袜制服| 大片电影免费在线观看免费| 婷婷色av中文字幕| 国产熟女欧美一区二区| 国产视频首页在线观看| 蜜桃久久精品国产亚洲av| 你懂的网址亚洲精品在线观看| 久久久国产一区二区| 国产中年淑女户外野战色| 欧美人与善性xxx| 26uuu在线亚洲综合色| 精品卡一卡二卡四卡免费| 午夜影院在线不卡| 欧美精品一区二区免费开放| 嫩草影院入口| 国产av一区二区精品久久| 国产精品福利在线免费观看| 一本久久精品| 国产午夜精品一二区理论片| 免费播放大片免费观看视频在线观看| 国产 精品1| 欧美人与善性xxx| 精品国产一区二区久久| 成年人午夜在线观看视频| 欧美老熟妇乱子伦牲交| 视频区图区小说| 亚洲欧美清纯卡通| 亚洲人成网站在线观看播放| 人人妻人人澡人人看| 亚洲av免费高清在线观看| 久久久欧美国产精品| 国产 一区精品| 午夜91福利影院| 欧美精品高潮呻吟av久久| 亚洲精品日韩在线中文字幕| 久久午夜福利片| 亚洲三级黄色毛片| 美女主播在线视频| 亚洲国产精品成人久久小说| 在线精品无人区一区二区三| 日韩制服骚丝袜av| 麻豆精品久久久久久蜜桃| 岛国毛片在线播放| 精品一区在线观看国产| 国产日韩欧美亚洲二区| 日本av免费视频播放| 亚洲高清免费不卡视频| 97超碰精品成人国产| 国产 一区精品| 一本久久精品| 国产精品国产av在线观看| 成人黄色视频免费在线看| 国产免费一区二区三区四区乱码| 午夜福利,免费看| 亚洲国产av新网站| 乱系列少妇在线播放| 欧美+日韩+精品| 久久久久久伊人网av| 人妻少妇偷人精品九色| 大香蕉久久网| 亚洲欧洲精品一区二区精品久久久 | 自拍欧美九色日韩亚洲蝌蚪91 | 精品国产一区二区久久| 国产黄色视频一区二区在线观看| 欧美最新免费一区二区三区| 中国国产av一级| 高清毛片免费看| 天天操日日干夜夜撸| 久久99热这里只频精品6学生| 涩涩av久久男人的天堂| 免费观看a级毛片全部| 2021少妇久久久久久久久久久| 另类亚洲欧美激情| 国产精品人妻久久久影院| 欧美丝袜亚洲另类| 男的添女的下面高潮视频| 久久青草综合色| av在线老鸭窝| 国产欧美日韩综合在线一区二区 | .国产精品久久| 黑人猛操日本美女一级片| 99热这里只有精品一区| 国产极品天堂在线| 欧美 亚洲 国产 日韩一| 在线播放无遮挡| 精品久久久久久电影网| 亚洲欧美日韩另类电影网站| 波野结衣二区三区在线| 9色porny在线观看| 大又大粗又爽又黄少妇毛片口| 街头女战士在线观看网站| 日韩欧美精品免费久久| 黄色配什么色好看| 亚洲av福利一区| 另类精品久久| 亚洲精品乱码久久久久久按摩| 啦啦啦视频在线资源免费观看| 成人综合一区亚洲| 亚洲美女视频黄频| 国产69精品久久久久777片| 欧美区成人在线视频| 日韩强制内射视频| 中文字幕久久专区| 我的女老师完整版在线观看| 韩国高清视频一区二区三区| av网站免费在线观看视频| 久久 成人 亚洲| 国产av一区二区精品久久| 国产精品秋霞免费鲁丝片| 亚洲精品成人av观看孕妇| 久久久久久久久久人人人人人人| 18禁在线播放成人免费| 99热这里只有是精品50| 亚洲伊人久久精品综合| 嘟嘟电影网在线观看| 纵有疾风起免费观看全集完整版| 精品国产国语对白av| 色5月婷婷丁香| 一本—道久久a久久精品蜜桃钙片| 建设人人有责人人尽责人人享有的| 国产av码专区亚洲av| 亚洲av电影在线观看一区二区三区| 一个人免费看片子| 亚洲国产av新网站| 纯流量卡能插随身wifi吗| 国产永久视频网站| 亚洲精品日本国产第一区| 国产精品.久久久| 国产精品欧美亚洲77777| 亚洲欧美成人精品一区二区| 99久久中文字幕三级久久日本| 国产精品国产三级国产专区5o| 在线天堂最新版资源| 日本爱情动作片www.在线观看| 亚洲丝袜综合中文字幕| 少妇人妻久久综合中文| 国产爽快片一区二区三区| 亚洲精品久久久久久婷婷小说| 午夜老司机福利剧场| 亚洲性久久影院| 国产国拍精品亚洲av在线观看| 三上悠亚av全集在线观看 | 高清毛片免费看| 亚洲真实伦在线观看| 一本大道久久a久久精品| 欧美激情国产日韩精品一区| 亚洲av男天堂| 免费黄网站久久成人精品| 日韩av在线免费看完整版不卡| 午夜福利影视在线免费观看| av天堂中文字幕网| 久久6这里有精品| 蜜桃在线观看..| 亚洲精品色激情综合| 亚洲精品国产av蜜桃| 午夜日本视频在线| 女性被躁到高潮视频| 国产精品一区二区在线观看99| 亚洲情色 制服丝袜| 男人添女人高潮全过程视频| 国产一区亚洲一区在线观看| 老女人水多毛片| 亚洲精品日韩av片在线观看| 麻豆精品久久久久久蜜桃| 永久免费av网站大全| 99久久综合免费| 精品少妇黑人巨大在线播放| 亚洲国产精品一区三区| 另类亚洲欧美激情| 男女国产视频网站| 成人午夜精彩视频在线观看| 国产亚洲欧美精品永久| 久久国产精品男人的天堂亚洲 | 日本av免费视频播放| 亚洲欧美一区二区三区国产| 91久久精品电影网| av女优亚洲男人天堂| 99视频精品全部免费 在线| 免费观看无遮挡的男女| 十八禁网站网址无遮挡 | 中文字幕久久专区| 街头女战士在线观看网站| 精品少妇内射三级| 一级毛片aaaaaa免费看小| 国产成人免费观看mmmm| 国产乱人偷精品视频| av免费观看日本| 日韩一区二区三区影片| 蜜臀久久99精品久久宅男| 简卡轻食公司| 人妻夜夜爽99麻豆av| 国模一区二区三区四区视频| h日本视频在线播放| 中国国产av一级| 日韩电影二区| av天堂中文字幕网| 国产成人a∨麻豆精品| 亚洲性久久影院| 欧美少妇被猛烈插入视频| 亚洲三级黄色毛片| 日韩av免费高清视频| 国产精品麻豆人妻色哟哟久久| 少妇人妻 视频| 亚洲成人一二三区av| 亚洲精品久久久久久婷婷小说| 黄色视频在线播放观看不卡| a级毛片在线看网站| 成年人午夜在线观看视频| 肉色欧美久久久久久久蜜桃| 最近手机中文字幕大全| 亚洲第一区二区三区不卡| 亚洲成人手机| 大陆偷拍与自拍| 国产精品久久久久久av不卡| 十八禁网站网址无遮挡 | 国产男女超爽视频在线观看| 99久久综合免费| 亚洲中文av在线| 校园人妻丝袜中文字幕| 看免费成人av毛片| 91精品国产国语对白视频| 人妻制服诱惑在线中文字幕| 国产高清有码在线观看视频| 五月开心婷婷网| av国产精品久久久久影院| 欧美人与善性xxx| 亚洲情色 制服丝袜| 久久亚洲国产成人精品v| 免费av中文字幕在线| 亚洲三级黄色毛片| 国国产精品蜜臀av免费| 久久久精品免费免费高清| 有码 亚洲区| 精品一品国产午夜福利视频| 有码 亚洲区| 日韩精品免费视频一区二区三区 | 欧美精品人与动牲交sv欧美| 欧美精品亚洲一区二区| 亚洲av中文av极速乱| 熟女人妻精品中文字幕| 国产黄色免费在线视频| 麻豆乱淫一区二区| 91aial.com中文字幕在线观看| 青春草国产在线视频| 午夜日本视频在线| 妹子高潮喷水视频| 免费观看a级毛片全部| 久热久热在线精品观看| 国产一区二区在线观看日韩| 老熟女久久久| 国产成人免费观看mmmm| 免费少妇av软件| 国产亚洲91精品色在线| 精品视频人人做人人爽| 亚洲欧洲精品一区二区精品久久久 | 久久精品国产a三级三级三级| 亚洲欧美成人综合另类久久久| 九九爱精品视频在线观看| 搡女人真爽免费视频火全软件| 亚洲精品国产av蜜桃| 国产极品粉嫩免费观看在线 | 夜夜爽夜夜爽视频| 97在线视频观看| 国产午夜精品一二区理论片| 一级毛片久久久久久久久女| 日本黄色日本黄色录像| 毛片一级片免费看久久久久| av天堂中文字幕网| 亚洲av免费高清在线观看| 人妻夜夜爽99麻豆av| 国产精品久久久久久精品电影小说| 中文天堂在线官网| 人妻少妇偷人精品九色| 一区二区三区精品91| 久久99热6这里只有精品| 日本av免费视频播放| 欧美成人精品欧美一级黄| 少妇丰满av| 国产深夜福利视频在线观看| 亚州av有码| 久久久国产欧美日韩av| 日本猛色少妇xxxxx猛交久久| 成年女人在线观看亚洲视频| 成人影院久久| 看免费成人av毛片| 欧美日韩国产mv在线观看视频| 高清午夜精品一区二区三区| 丝袜脚勾引网站| 国产永久视频网站| 啦啦啦啦在线视频资源| 26uuu在线亚洲综合色| 乱人伦中国视频| 国产精品国产三级国产av玫瑰| 少妇精品久久久久久久| 一级a做视频免费观看| 18+在线观看网站| 青春草亚洲视频在线观看| 亚洲在久久综合| 黄色毛片三级朝国网站 | 最近的中文字幕免费完整| 亚洲精品自拍成人| 超碰97精品在线观看| 一区在线观看完整版| 精品99又大又爽又粗少妇毛片| 少妇人妻 视频| 久久久国产欧美日韩av| 97在线视频观看| 日日摸夜夜添夜夜添av毛片| 91精品国产国语对白视频| 最近中文字幕2019免费版| 大香蕉97超碰在线| 国产色婷婷99| 国产亚洲最大av| 伦精品一区二区三区| 黄色毛片三级朝国网站 | 亚洲精品成人av观看孕妇| 国产精品无大码| 久久久久久久国产电影| 亚州av有码| 日韩欧美一区视频在线观看 | 亚洲经典国产精华液单| 欧美 亚洲 国产 日韩一| 国内少妇人妻偷人精品xxx网站| a级毛色黄片| 少妇被粗大猛烈的视频| 国产高清三级在线| 欧美国产精品一级二级三级 | 亚洲欧美清纯卡通| 日韩av在线免费看完整版不卡| 啦啦啦啦在线视频资源| 亚洲伊人久久精品综合| h视频一区二区三区| 偷拍熟女少妇极品色| 国产一区二区三区综合在线观看 | 亚洲色图综合在线观看| 性高湖久久久久久久久免费观看| 春色校园在线视频观看| freevideosex欧美| 蜜桃久久精品国产亚洲av| 黑人高潮一二区| 人妻 亚洲 视频| 黄色怎么调成土黄色| av天堂中文字幕网| 精品人妻熟女av久视频| 欧美日韩视频精品一区| 日日啪夜夜爽| 欧美另类一区| 国产成人精品福利久久| 一本一本综合久久| 日产精品乱码卡一卡2卡三| 午夜精品国产一区二区电影| 亚洲激情五月婷婷啪啪| 亚洲精品成人av观看孕妇| 欧美日韩精品成人综合77777| 中文在线观看免费www的网站| 精品久久国产蜜桃| 成年人免费黄色播放视频 | 国产白丝娇喘喷水9色精品| 亚洲精品国产av成人精品| 久久久久久久久久久久大奶| 国产精品伦人一区二区| 日日撸夜夜添| 国产精品人妻久久久久久| 日日摸夜夜添夜夜爱| 亚洲欧美一区二区三区黑人 | 久久久久久久久大av| 99久久精品一区二区三区| 久久99热这里只频精品6学生| 人体艺术视频欧美日本| 午夜激情福利司机影院| 精品国产乱码久久久久久小说| 99热这里只有是精品50| 亚洲情色 制服丝袜| 黑丝袜美女国产一区| 国产精品99久久99久久久不卡 | 国产免费视频播放在线视频| 国产免费一级a男人的天堂| 热re99久久精品国产66热6| 亚洲美女视频黄频| 久热这里只有精品99| 一区二区三区四区激情视频| 成年美女黄网站色视频大全免费 | 亚洲无线观看免费| 一级毛片我不卡| 婷婷色麻豆天堂久久| 在线亚洲精品国产二区图片欧美 | 亚洲国产精品成人久久小说| 精华霜和精华液先用哪个| tube8黄色片| 偷拍熟女少妇极品色| 人体艺术视频欧美日本| 国产乱来视频区| 少妇人妻久久综合中文| 18+在线观看网站| 亚洲高清免费不卡视频| 亚洲精品自拍成人| 国产在线一区二区三区精| 免费av中文字幕在线| 女性生殖器流出的白浆| 日韩 亚洲 欧美在线| 欧美bdsm另类| 日韩一区二区视频免费看| 亚洲欧美日韩卡通动漫| 三级国产精品欧美在线观看| 国产精品蜜桃在线观看| 国产亚洲欧美精品永久| 黄色一级大片看看| 熟女人妻精品中文字幕| 精品国产一区二区久久| 人人妻人人看人人澡| 美女福利国产在线| 日韩av在线免费看完整版不卡| 九九爱精品视频在线观看| 国国产精品蜜臀av免费| 久久女婷五月综合色啪小说| 久久久久国产精品人妻一区二区| 菩萨蛮人人尽说江南好唐韦庄| 赤兔流量卡办理| 好男人视频免费观看在线| 亚洲av中文av极速乱| 精品亚洲乱码少妇综合久久| 在线观看国产h片| 成人亚洲欧美一区二区av| 少妇人妻精品综合一区二区| 日本vs欧美在线观看视频 | 99九九在线精品视频 | 亚洲不卡免费看| 激情五月婷婷亚洲| 波野结衣二区三区在线| 麻豆乱淫一区二区| 亚洲精品乱久久久久久| 精品人妻熟女毛片av久久网站| 国产精品嫩草影院av在线观看| 91aial.com中文字幕在线观看| 国产高清三级在线| 久久青草综合色| 日韩欧美 国产精品| 亚洲国产成人一精品久久久| 亚洲欧洲日产国产| 久久国产精品大桥未久av | 国产精品秋霞免费鲁丝片| 久久久欧美国产精品| 午夜老司机福利剧场| av播播在线观看一区| av在线老鸭窝| av国产精品久久久久影院| 美女大奶头黄色视频| 亚洲精品乱码久久久v下载方式| 亚洲欧洲精品一区二区精品久久久 | 欧美人与善性xxx| 五月天丁香电影| 又大又黄又爽视频免费| 18+在线观看网站| 春色校园在线视频观看| av免费观看日本| 99久久精品一区二区三区| av国产久精品久网站免费入址| 自拍偷自拍亚洲精品老妇| 成人美女网站在线观看视频| 啦啦啦视频在线资源免费观看| 91午夜精品亚洲一区二区三区| 极品少妇高潮喷水抽搐| 日韩成人av中文字幕在线观看| 午夜福利网站1000一区二区三区| 午夜精品国产一区二区电影| 精品一区在线观看国产| 欧美日韩一区二区视频在线观看视频在线| 嘟嘟电影网在线观看| 国产亚洲午夜精品一区二区久久| 国产高清不卡午夜福利| 国产精品久久久久久精品古装| 久久久国产一区二区| 久久女婷五月综合色啪小说| 久久精品久久久久久久性| 又粗又硬又长又爽又黄的视频| 女性生殖器流出的白浆| 三级国产精品片| 午夜激情久久久久久久| 久久久久久久久久久免费av| 天堂俺去俺来也www色官网| 又粗又硬又长又爽又黄的视频| 国产黄片美女视频| 丰满饥渴人妻一区二区三| 国产国拍精品亚洲av在线观看| 欧美最新免费一区二区三区| 超碰97精品在线观看| 国产亚洲午夜精品一区二区久久| 九九在线视频观看精品| 韩国高清视频一区二区三区| 国产亚洲欧美精品永久| 黑丝袜美女国产一区| 国产精品久久久久久久久免| 国产精品三级大全| 国产有黄有色有爽视频| 晚上一个人看的免费电影| 亚洲欧洲精品一区二区精品久久久 | 日韩熟女老妇一区二区性免费视频|