• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    N-fold Bernoulli probability based adaptive fast-tracking algorithm and its application to autonomous aerial refuelling

    2023-02-09 09:00:20JarhinbekRASOLYueleiXUQingZHOUTianHUIZhaoxiangZHANG
    CHINESE JOURNAL OF AERONAUTICS 2023年1期

    Jarhinbek RASOL, Yuelei XU, Qing ZHOU, Tian HUI, Zhaoxiang ZHANG

    Unmanned System Research Institute, Northwestern Polytechnical University, Xi’an 710072, China

    KEYWORDS Autonomous aerial refueling;N-fold Bernoulli probability theorem;Object detection;Object tracking;YOLOv4

    Abstract Recently,deep learning has been widely utilized for object tracking tasks.However,deep learning encounters limits in tasks such as Autonomous Aerial Refueling (AAR), where the target object can vary substantially in size, requiring high-precision real-time performance in embedded systems. This paper presents a novel embedded adaptiveness single-object tracking framework based on an improved YOLOv4 detection approach and an n-fold Bernoulli probability theorem.First, an Asymmetric Convolutional Network (ACNet) and dense blocks are combined with the YOLOv4 architecture to detect small objects with high precision when similar objects are in the background.The prior object information,such as its location in the previous frame and its speed,is utilized to adaptively track objects of various sizes.Moreover,based on the n-fold Bernoulli probability theorem,we develop a filter that uses statistical laws to reduce the false positive rate of object tracking.To evaluate the efficiency of our algorithm,a new AAR dataset is collected,and extensive AAR detection and tracking experiments are performed. The results demonstrate that our improved detection algorithm is better than the original YOLOv4 algorithm on small and similar object detection tasks; the object tracking algorithm is better than state-of-the-art object tracking algorithms on refueling drogue tracking tasks.

    1. Introduction

    The use of Autonomous Aerial Refueling (AAR) can increase the range of an Unmanned Aerial Vehicle(UAV).1It has high application value in military and civilian fields. There are two main types of aerial refueling: Flying Boom Refueling (FBR)and Probe-Drogue Refueling (PDR). Compared with FBR,PDR meets the requirements of UAVs,such as high flexibility,high safety, and simplicity. Thus, PDR is more suitable than FBR for unmanned aerial systems, and this paper is oriented to the PDR-based AAR task.

    For the AAR task, the refueling drogue must be detected correctly. Then, an object tracking algorithm is required to accelerate the proceeding speed for real-time detection.Generally,object tracking tasks consist of image processing,machine learning,and optimization,and they are the premise and foundation of the AAR task.2,3

    The task of refueling drogue detection and tracking can be divided into two types: active vision methods and passive vision methods.4The former uses artificial features such as painted markers5-8and light emitting diode beacons9,10to detect and track refueling drogues. Although the active vision method has the advantages of high speed and high reliability,the disadvantages are also obvious.It needs to modify the tanker, for example by installing LEDs, which need additional power,and such modification is apt to malfunction when used in real-world situations. The passive vision method mainly relies on intrinsic vision features to detect refueling drogues.Passive vision methods can be divided into two types: nondeep learning based tracking algorithm11-13and deep learning based tracking algorithm.11,14-16Although the non-deep learning based algorithm has high speed and can satisfy the realtime requirement, the robustness is not sufficient to apply it to real-world tasks.

    The deep learning based object tracking algorithm uses the powerful presenting ability to track objects. Many deep learning based object tracking algorithms have been proposed to improve the tracking performance. The MDNet17tracking algorithm designs a small network to learn features, uses softmax18to classify samples, and has high performance, except that its speed is only one Frame Per Second(FPS).The SiamFC algorithm19uses a Siamese network to train a similar function offline,which can be used to select the most similar candidate in each frame of video.The Siamese instance search network algorithm20directly uses the Siamese network to learn the matching function of object and candidate templates.It uses the object in the first frame only as a template to track the object during the tracking step. Based on the Siamese network, it uses a region proposal network to directly estimate object scales,21which promotes tracking performance and efficiency. The DaSiamRPN22tracker introduces a distractor-aware model to increase accuracy further. Furthermore, SiamRPN++23uses ResNet and Inception to replace the baseline network,which increases the robustness of the tracking algorithm.

    Although all of these object tracking algorithms can be used to track refueling drogue, they cannot simultaneously handle the situation encountered in refueling drogue tracking task,in which(A)the target object has a large range of scales,(B)the target object is sometimes too small, and(C)high precision with real-time operating speed is needed in an embedded system. To address these problems, we develop an adaptive tracking algorithm, as shown in Fig. 1. We improve the YOLOv424by introducing the dense block and ACNet,25which enhances the ability to detect small objects and increases the precision with no computational complexity added in the inference stage. We also use the n-fold Bernoulli probability theorem to design a filter that can screen out false positive samples when tracking an object. When we combine this filter with improved YOLOv4, the detecting precision will be high enough to satisfy the AAR task. Finally, we develop a tracking-by-detection algorithm based on this improved YOLOv4 and the filter mentioned above for visually tracking refueling drogues in AAR tasks. The main contributions of this article are as follows:

    (1) In this paper,we improve YOLOv4 by(A)replacing the residual block with a dense block to increase the feature extraction capability,which enhances the detection ability on small objects.(B)The ACNet25theorem is used to enrich the network’s receptive field by putting differentsized kernels parallel to the 3 × 3 kernel. This increases the precision when similar objects are in the background,but does not increase the computational burden in the inference stage.

    (2) We propose an adaptive tracking-by-detection algorithm based on the improved YOLOv4 for the AAR task. The proposed algorithm utilizes the prior knowledge of the object, such as the object’s location, moving speed and direction, to adaptively track the object,which may vary in size, with the real-time operating speed needed in an embedded system.

    (3) Based on the n-fold Bernoulli probability theorem, we develop a filter that synthesizes the previous information of the object to decrease the false positive rate, which can increase precision.

    In Fig.1,the dashed rectangles correspond to our contributions. The dashed arrow line represents the improvement that we apply to the YOLOv4 algorithm.

    The remainder of this paper is organized as follows: Section 2 introduces the improvement applied to YOLOv4. The object tracking algorithm that we design is described in Section 3.Section 4 presents the experimental results and analysis,and finally, Section 5 presents concluding remarks.

    2. Improvement on YOLOv4

    YOLOv4 is an object detection algorithm based on the YOLO algorithm. It uses the cross-stage partial 53-layer Darknet(CSPDarknet53) as the backbone network for feature extraction. CSPDarknet53 enriches the network gradient combination by introducing Cross-Stage Partial connections (CSPs)while reducing the amount of calculation.Therefore,YOLOv4 has better performance and is faster than other object detection algorithms.

    2.1. Enhancing feature extraction ability of network

    CSPDarknet53 in YOLOv4 is constructed with several CSP-n blocks,and a CSP-n block mainly consists of n residual blocks that connect back and forth layers by simply adding their feature maps together. Although this structure can make the training process easier, it limits the feature extraction capability of the network by assigning the same weights to all feature maps added together in one block, which consequently restricts the network’s detecting capability on small objects.On the other hand, the dense block26assigns different weights to different feature maps concatenated together in one block,enhancing the network’s feature extraction ability.

    Fig. 1 Aerial refueling drogue tracking algorithm.

    Therefore, in this paper, we replace every residual block in all CSP-n blocks with dense blocks to improve the information flow between layers, as shown in Fig. 2 (in Fig. 2, we take the CSP-2 block for illustration). The dense block consists of a 3 × 3 convolutional layer followed by a 1 × 1 convolutional layer. In each dense block (×n), every 3 × 3 convolutional layer directly connects to every 1 × 1 convolutional layer in the following blocks by concatenation, as shown in the solidline box in Fig. 2. The 1 × 1 convolutional layers are used to merge the feature maps produced by each 3 × 3 convolutional layer of the front block.

    Fig. 2 Application of dense block.

    The CSPDarknet53 structure proves to have powerful feature extraction capabilities. Using dense blocks to replace the residual blocks does not change the overall structure; instead,only the pointwise addition operator needs to be replaced with the concatenation operator. The dense block has a more powerful feature representing ability than the residual block.Therefore, using the dense block to replace the residual block can improve the feature extracting ability of CSPDarknet53.

    In Fig. 2, CBA represents convolutional, normalization,and activation layers that are serially connected in sequence.The ‘‘-n” is the number of dense blocks or residual blocks,e.g., in Fig. 2, n is set to 2. ‘‘s1” and ‘‘s2” represent the stride values of the convolutional layer.

    2.2. Enriching scales of receptive fields

    For an object detection algorithm, the receptive field heavily influences the semantic features that the detection algorithm extracts,which will directly affect the detection accuracy when there is a similar object in the background,and the object has a wide range of scales. The YOLOv4 network consists of only 3×3 and 1×1 convolutional layers,which limits the receptive field to a square shape.Consequently,this will restrict the representation of semantic features. To address this problem, we must make the receptive field have more scales.Most methods do this by introducing architectural changes to the network,which will introduce an additional computational burden at the inference stage. Although our strategy brings architectural change to the network, by introducing the theory of ACNet,the computational burden exists only in the training process,not in the inference process. Therefore, there is no extra computational burden on the network.

    The main theory of ACNet is described as follows:if several 2D kernels have compatible sizes,they work on the same input with the same stride to produce the output with the same size,and their outputs are directly summed.Then,the same output can be achieved by adding these kernels to the corresponding positions. For convenience, we marked this property as the Additivity Property of Convolution (APC). Compatibility means that the smaller kernel can patch into larger kernels,i.e.,the 1 × 3, 1 × 1, and 3 × 1 kernels are compatible with 3 × 3 kernels.Formally,if condition(1)is satisfiable on layers q and p, transformation (2) can be applied to layers q and p.

    where Mand Mare the inputs of the q-th and p-th layers,respectively;Y and X are the dimensions of the kernel;D is the filter number of the convolutional layer;E is a matrix;K(1)and K(2)are two compatible 2D kernels; * and ⊕are a convolutional operation and elementwise addition of the kernel parameters on the corresponding position, respectively.

    Usually, networks adopt a batch normalization layer to accelerate the training process and enhance the feature representational power. To apply the APC to the networks that use batch normalization after the convolutional layer, ACNet fuses Batch Normalization (BN)into the convolutional kernel parameters. Afterward, the kernels fused with BN can be directly fused into one square kernel using APC.

    To widen the receptive field scales of the network, during the training stage, we change each of 3 × 3 convolutional layers and the corresponding BN layer in YOLOv4 into 3 × 3,3 × 1, and 1 × 3 convolutional layers with the corresponding BN,as shown in the purple dashed box on the left-hand side of Fig. 3. The three branches have identical numbers of filters,output resolutions, and numbers of output channels. Then,we fuse the BN in each branch into a convolutional kernel using ACNet BN fusion theory. Then the APC is applied to fuse the branches into one standard convolutional layer by adding the three kernels onto the corresponding positions of the square kernel.Finally,we use the fused weights to initialize the original network—the YOLOv4 after using the dense block to replace the residual block in the last section,as shown in the right part of Fig.3.Note that the original network mentioned in the last sentence has no BN after each 3 × 3 convolutional layer. This is because the BN has been fused into each 3 × 3 convolutional kernel parameter.

    By using ACNet, we expand the scale of the receptive field to a certain extent, thereby improving the detection accuracy of the model. Compared to other methods, this method does not introduce additional operations during the inference stage.

    3. N-fold Bernoulli probability theorem based adaptive fasttracking algorithm

    3.1. Adaptive fast tracking

    If the tracking object is certain, we can use a modified object detecting algorithm to track it. It is more reliable than other object tracking algorithms to use the object detection algorithm to track a specific object because the object detection algorithm has prior knowledge about the tracking object.However, it is well known that deep learning based object detection algorithms have very high complexity, which limits their use in object tracking tasks in an embedded device. To address this issue, we develop a tracking-by-detection algorithm based on YOLOv4,which has higher precision and more robustness than an arbitrary object tracking algorithm. Moreover, it has a high inference speed that ensures its use in an embedded device.

    In YOLOv4, the input can have different sizes during online use. Decreasing the input size can significantly accelerate the inference speed of the algorithm. For example, given that the original input size is H×W and the decreased input size is H'×W', the computational complexity decreases as follows:

    where R denotes the decreasing ratio of complexity. Usually,the input shape of a network is square. Let I denote the original input size of a network, and I'denote the decreased input size of the network. Eq. (3) is transformed to.

    Fig. 3 Application of ACNet.

    To benefit from this property for real-time applications,we need to greatly reduce the input size.However,if we carelessly reduce the input size, the object size becomes too small to be detected when the object size is not large enough in the original image.However,if we know the object’s approximate location in an image, we can crop the image around the object so that the cropped area is much smaller than the area of the original image. The object detection algorithm itself can provide the object location in the first frame. The rest of the object’s approximate location can be given by the former frame because the objects in two adjacent frames are close to each other. Based on this, we need to crop a larger area around the bounding box of the former frame and feed it into the network to detect the object in the cropped image.Of note is that the object tracking algorithm and the object detection algorithm are identical except for the input size. Thus, the smaller input size network has lower computational complexity, as shown in Eq. (4).

    Here, the input size is not equal to the cropping size. The input size of YOLOv4 should be a multiple of 32, and when used to track the refueling drogue, the input size of YOLOv4 needs to be sufficiently small for real-time application. The cropping size needs to be considered comprehensively to ensure that (A) the size of the object is sufficiently large for the network to detect,and that(B)the object is in the cropped image. Condition (A) is called the upper bound condition because for a certain sized object, a larger image corresponds to a smaller object size in the input image(because the cropped images are resized back to the same input size). Condition (B)is called the lower bound condition because the smaller cropped size corresponds to narrower vision. Therefore, we need to determine the upper and lower bounds of the cropping size. For clarity, we illustrate some of the variables in Fig. 4.

    (1) Upper bound calculation

    The object detection network has limits for objects that are too small in the image.If the object’s size in the image is smaller than this limit size, the network is blind to that object. Let SLdenote the limit size, Sobthe object’s original size in an image, and WCand HCthe cropped width and height, respectively.Then,to ensure that the object size in the input image is larger than the limit size, WCand HCmust satisfy.

    By satisfying Eq. (6), the object size in the input image becomes larger than SL, which can ensure that the detection algorithm can detect the tracking object.Usually, the cropped image size is larger than the input size of the network.We need to resize back the cropped image to match the input size of the network. Therefore, if SCis larger than this upper boundary when we resize back the cropped image, the object size in the input image will be smaller than the limit size.

    (2) Lower bound calculation

    Algorithm 1. Adaptive tacking 1. Initialize r to 1 2. for every frame 3. for index, ξindex in ξ 4. if ξindex+1 >Sob×r >ξindex 5. e = index + 1 6. break 7. else 8. continue 9. end 10. end 11. SC =I ×2e 12. if SC does not satisfy Eq. (6)13. return SC to the previous value 14. end 15. if this is not the first frame 16. let d ≈Li - Li+1 17. vcam = d × f 18. end 19. if SC does not satisfy Eq. (9)20. let SC =2×vcam f 21. end 22. r=I'SC 23. resize cropped image to I' × I'24. execute detection 25. update Sob, Li 26. end

    Fig. 4 Illustration of variables.

    Given that the size of the object that needs to be tracked is SRobin the moving direction and the moving speed is v(relative to the camera), we can calculate the field of view l as follows:In Algorithm 1,the threshold list ξ is sorted from the smallest to the largest. r is the resizing ratio to resize cropped images back to the input size. The larger the cropped size, the larger the field of view and the more robust the algorithm.Therefore,on line 11,we enlarge the cropped size to obtain as large a field of view as possible. In this paper, we modify the crop size by doubling or halving each time it needs to be changed,as shown on line 11. The scaling parameters e on line 11 is set to index + 1, as shown on line 5. However, it must satisfy Eq.(6) to ensure that the object’s size in the input image is larger than the limit size. Otherwise, the algorithm is blind to the object, which is ensured on line 12. However, in some cases,Eq. (6) and Eq. (9) cannot both be satisfied. In that case, we choose to compromise Eq. (6) because if Eq. (6) is violated,there is a certain probability that the target can be detected,but if Eq.(9)is violated,it will be impossible to detect the target object because the object will no longer be in the input image.

    It should be considered that the tracking algorithm fails to track the object in a certain frame.The algorithm used to track the object is based on YOLOv4, which can change the input size when using it online. Therefore, if the tracking algorithm loses the object, we can feed the whole image to the detection network with an input size of 608 × 608 to find the object.After finding the object, we can switch back to the tracking algorithm.

    3.2. N-fold Bernoulli probability theorem based filter

    For a binary classification task,we assume that:Ak=‘‘k samples are predicted as positive,” k = 0,1,···,n, and B = ‘‘the samples predicted as positive are more than 50% of the total predicted samples”. Clearly, the events that the samples are predicted as positive or negative are independent and identically distributed events. Thus, if we set the detecting precision as p, according to the n-fold Bernoulli probability theorem:

    To analyze the relationship among P(B),p,and n,we take p from 0 to 1 with intervals of 0.05. Then, for each p, we calculate P(B)based on n,which is taken from 2 to 50 with intervals of 1, and draw them in Fig. 5. In Fig. 5, the vertical and horizontal coordinates are the dependent variable P(B) and independent variable n in Eq. (14), respectively.

    From Fig. 5(a), when p is greater than 0.5, it rises in a zigzag shape with increasing n;when p is less than 0.5,it falls in a zigzag shape. From Fig. 5(c), when n is an odd number, P(B)monotonically increases with p and becomes greater than p when p is larger than 0.5; this trend reverses when p is smaller than 0.5. This is not true for even numbers n, as shown in Fig. 5(b).

    In a single-object detection task, we can regard the detection of an object from the background as a binary classification task. The challenge is to ensure that the object in one frame is identical to the object in the former frame. For this issue, we first use the former frames to estimate the speed and direction of the object,and if there is only one object near the estimated location, we simply regard it as the tracking object. Suppose that there is more than one object around the estimated location.In that case,we use the Structural Similarity Index Measure (SSIM)27to calculate the similarity of the image in the object bounding box with the former one and choose the bounding box with the highest score as the final object bounding box.

    Moreover, according to Fig. 5, this process enhances the part above 0.5 and can depress the part below 0.5. Therefore,after ensuring the target object, we can regard the rest of the image as negative(nonobject),which can filter out noisy detection results in some frames. Based on the above analysis, we develop a filter to diminish false positive results, which consequently improves the tracking precision. The algorithm is shown in Algorithm 2.

    In Algorithm 2, n represents n in Eq. (14), which is usually set to an odd number according to the above analysis. The update on line 24 pops the first element from the queue Zjand appends the new result. If the results in the queue Zjare less than n,we must first fill up the queue.Line 16 corresponds to P(B) in Eq. (14), and the real object on line 17 means that there is a possible object in the cropped image. Because the AAR task is a single object tracking task, we need to select only one object in the end. Therefore, if more than one object are in the cropped image, we use the confidence to select the tracking object on line 27.

    Fig. 5 Relationship among P(B), p, and n in Eq. (14).

    This tracking algorithm can also be used for other single object tracking tasks. However, to get better performance,some modifications are needed. The threshold list ξ should be experimentally chosen according to the limit size in Eq.(6) and the object’s size. Given a certain limit size, we need to choose a proper threshold list to ensure that the object exists in the cropped image. Instinctively, the smaller the threshold is, the smaller the probability that the object will exceed the cropped image is. However, if the object size is not large enough when enlarging the cropping size, the object size will be smaller than the limit size, leading to missing detection when tracking the object. On the other hand, if the values in the threshold list are too large, the probability of the object being outside of the cropped image increases.

    There are other limits to using the algorithm in other fields.The algorithm proposed in this paper is based on an object detection algorithm, which cannot distinguish objects belonging to the same class. Therefore, the proposed algorithm cannot be applied in scenarios with many objects belonging to the same class in the background,such as tracking a person in the street,tracking birds in a flock,etc.However,it can be used in the AAR task because the probability that another refueling drogue will exist in the receiver’s field of vision is remote.

    Algorithm 2.N-fold Bernoulli probability theorem based filter 1. Maintain a list that contains the detected object Q= Q0,Q1,···[ ]2. For every object Qj maintains a queue of size n, which holds the results Zj = Z1,Z2,···,Zn[]3. for every frame 4. if the algorithm detects a new object 5. add that object to the list Q 6. end 7. for every object Qj 8. estimate the object location Li =vcam f +Li-1 9. if the location Li is outside the cropping range 10. Delete Qj from the object list 11. return to line 7 12. end 13. if there is more than one object near the estimated location 14. Use SSIM to select the most similar one as an object Qj 15. end 16. if the positive result is more than half in queue Zj 17. Regard it as the real object 18. else 19. Regard it as background 20. end 21. if there are n negative results in the queue Zj 22. Delete Qj 23. else 24. Update Zj 25. end 26. if there is more than one real object 27. Select the one with the highest confidence as the tracking object 28. else 29. regard the real object as the tracking object 30. end 31. end

    4. Experiment

    In this section, we first evaluate our improvement on YOLOv4, and compare our algorithm with other state-ofthe-art algorithms: YOLOv5L28(YV5L) and EfficientDet29(EFD). Then, we test the influence of the input size on the tracking algorithm and verify the effectiveness of the n-fold Bernoulli probability theorem based filter. Finally, we compare our tracking algorithm with other tracking algorithms:Efficient Convolution Operators30(ECO), DaSiamRPN,22MDNet,17and SiamRPN++.23

    4.1. Dataset

    For the dataset,we make a refueling drogue at a 1:1 scale and take many videos using this refueling drogue.These videos are mainly taken from distances of 2 to 40 m. Therefore, these videos cover the size of the refueling drogue in an image from approximately two meters to approximately forty meters. The object size distribution in the dataset is shown in Fig. 6. To simulate a practical situation, we download many refueling videos from the Internet. Afterward, we choose 33 videos out of 37 videos, split the video frames, randomly select 14700 images from the video frames,and set them as the training set.For the test set,because we need to verify the detection algorithm and tracking algorithm, we label every frame of the remaining videos and sequentially store each video frame in different folders.

    4.2. Implementation details

    During the training process, we train our algorithm on the training dataset from scratch. We use mosaic, random rotations, Gaussian noise, and Gamma transfer augmentation.The Adam weight decay optimizer is used to optimize the network; the learning rate is set to 10-3; the batch size and minibatch size are set to 130 and 26, respectively. We iterate the network for 250 epochs on the dataset.

    During the tracking process, we set the input size of our tracking algorithm to 128 × 128 pixels, n in Eq. (14) to 9,and ξ in Algorithm 1 to list(60,120,200,350,400).The cropping area adaptively changes according to the size of the object in Algorithm 1.

    The algorithm is programmed based on PyTorch 1.6 and carried out on an NVIDIA TITAN RTX GPU,and we implement it on a Jetson AGX Xavier to verify its efficiency on the embedded device.

    4.3. Evaluation of improved YOLOv4

    After setting up the operating environment, we compare our improved YOLOv4 with the original YOLOv4 on the test dataset by drawing the PR curve. When calculating the precision and recall, we regard the samples that box only the refueling drogue as True Positive (TP) instances; the samples that box something else are regarded as False Positive (FP)instances; the samples that miss the refueling drogue are considered False Negative(FN)instances.The equations of precision and recall are:

    Fig. 6 Object size distribution in training dataset.

    In the test, we use Dense-CSPDarknet (DC) to refer to the strategy that uses a dense block to replace the residual block.ACNet means that the 3 × 3 layers are replaced by placing 1×3 and 3×1 layers parallel to the 3×3 layers at the training stage and setting them back to 3×3 layers at the inference stage using ACNet theory. Ultimately, we obtain Table 1.

    To ensure that our strategy positively affects object detection, we separately test every proposed strategy in this paper.We list the tested algorithms in Table 1 and plot the testing results in Fig. 7. To better examine the results, we enlarge the upper right corner of Fig. 7(a). The experimental results show that ACNet increases the precision,and the DC strategy increases the recall ratio. The reason is that the ACNet strategy can widen the range of the receptive field scale to represent more semantic information, and dense blocks have a more powerful feature extraction capability,which can help the network extract detailed information to distinguish a small object from the background. Thus, YV4_D_A has higher precision and recall ratio than the other methods.

    We also visualize some of the samples with different prediction results on these algorithms. The visualized samples are shown in Fig. 8. In Fig. 8, the three columns correspond to the results of the three algorithms. For clarity, we enlarge the content in the yellow box area. In the first row of Fig. 8,YV4 did not find the target, while YV4_D found the target object, but it also recognized something else as the target object. Only the YV4_D_A gave the correct result. In the second row, although YV4 found the object, it also regarded the engine as the drogue. YV4_D and YV4_D_A correctly recognized the drogue.Although they all found the correct object in the last row,YV4 and YV4_D had false detections for the last row. Only the YV4_D_A correctly found the drogue again.From the above description, YV4_D has a stronger detection capability for small objects than YV4, but the accuracy is not as high as YV4_D_A. Additionally, YV4_D_A has a higher confidence in predicting the target object.

    For a more detailed comparison, we separate the test dataset into three scales: small, normal, and large. The size of a small object is smaller than 50 pixels, the normal object is greater than 50 pixels and less than 300 pixels, and the large object is greater than 300 pixels. YV5L and our methods are tested on three different input sizes, 416, 544, and 640. For the EFD, we implement the three fastest versions of the EFD, i.e., D0, D1, and D2. All algorithms are tested with batch size 8.We draw our testing result in Fig.9.As seen fromFig. 9, although YV5L is faster than our algorithm and has similar performance on normal and large objects, its performance on small targets is not as good as our algorithm. The performance on small objects can enable the aircraft to accurately identify the refueling drogue farther away from the refueling drogue to buy more time for the aircraft to adjust. EFD cannot satisfy the real-time requirement of the AAR task, the fastest version of EFD achieves only 62 FPS on a TITAN RTX, and the Average Precision (AP) is below that of our algorithm.

    Table 1 Algorithms to evaluate efficacy of our algorithm.

    Fig. 9 Comparison of object detection algorithms.

    4.4. Evaluation of tracking algorithm

    To demonstrate the advantage of our tracking algorithm in the autonomous aerial refueling task, we compare our algorithm with some state-of-the-art algorithms on the test dataset. In this paper,we use a precision plot and success plot to evaluate the algorithms, which is a common benchmark of tracking algorithms. During the tracking algorithm test, we add the reversed sequence video to the test dataset.

    Fig. 8 Visualized samples with different results on different algorithms.

    (1) Precision plot: Percentages of frames whose estimated locations lie in a given threshold distance to the ground truth centers.

    (2) Success plot: A frame whose overlap is larger than the threshold is termed a successful frame, and the ratios of successful frames at thresholds ranging from 0 to 1 are plotted in the success plots.

    Fig. 10 Object size distribution and recall and precision of different object-sized datasets.

    First,we test the influence of the cropping size on the tracking performance. To accomplish this, we first need to determine the limit size of our detection algorithm. To find the limit size, we set multiple continuous size intervals and divide the dataset into multiple sets according to the interval that contains the object size.Fig.10(a)and Fig.10(b)show the distribution of widths and heights.The different colors in Fig.10(a) and Fig. 10 (b) represent different object-sized datasets.The algorithm is tested separately on these sets. When there is a considerable drop in the recall at a certain interval,we consider the upper bound of the interval to be the limit size. We record the precision and recall of the algorithm on each interval set, and the results are shown in Fig. 10 (c). In Fig. 10 (c),we omit the results achieved on the datasets in which the object sizes are larger than 70 pixels. This is because the recall rate already achieves stability when the object size reaches 70 pixels. It can be seen from Fig. 10 (c) that there is a great loss in recall rate when the size passes 26 pixels. Accordingly, we choose 26 as the limit size of our algorithm, which means that when the object in the input image is smaller than 26 pixels,the algorithm will have a difficulty finding it in images.Afterward,we decrease the input size of the detection algorithm to every size divisible by 32 from 224 pixels to 64 pixels and record the missing frames and operating time on the embedded system. The results are shown in Table 2.

    Table 2 shows that the input size of the network greatly affects the operating speed because the input size of the network can change the computational complexity of the network, as shown in Eq. (4). From a speed perspective, the smaller the input size, the faster the network. In contrast, the effect of the input size on accuracy is the opposite of the speed.This is because when the cropping size is not sufficiently large to contain the object in the cropped image,we need to enlarge the cropping size to ensure that the object exists in the cropped image. Afterward, we must resize the cropped image to fit the input size. However, if the cropping size is much larger than the input size, many object details are lost when resizing the image back to the input size, making the network extract blurred features. Thus, the result of the network is based on blurred features, which decreases the performance. In general,the cropping size cannot be smaller than a certain size,although a smaller size corresponds to a higher speed. Therefore, we need to make a tradeoff between precision and operating speed.In this paper,we choose a size of 128 pixels for the AAR drogue tracking task.

    Second, we test the validity of the n-fold Bernoulli probability theorem based filter. We test the algorithms (with and without the n-fold Bernoulli probability theorem based filter)on all testing datasets. In Eq. (14), n can be set to different numbers.According to our analysis,if n is set to an odd number, it can increase the result when p is larger than 0.5 and decrease it when p is smaller than 0.5.The larger n is,the better the results are.In our test,n is set to odd numbers from 3 to 11,and‘‘-n”is used to indicate the chosen number.The results are listed in Table 3. In Table 3, ‘‘Proposed-n” indicates the YV4_D_A algorithm with an n-fold Bernoulli event algorithm.Table 3 shows that the precision increases with increasing n,while the recall remains the same, because we can only weed out the FP results using the former information of the object but cannot decrease the FN results.

    Table 2 Performance of network with different input sizes.

    Finally,we compare our algorithm on an embedded system with other state-of-the-art object tracking algorithms. We first use the improved YV4_D_A to detect refueling drogues and submit the drogue locations to tracking algorithms. We use the same strategy to test all algorithms on the test data.In this test,we use the MobileNetv2 backbone version of SiamRPN++,23and the template and input sizes are set to 127 and 255,respectively. For the SiamAPN++,31we set the input size and template size to 127 × 127 and 287 × 287, respectively.In ECO,30we set the maximum number of the components to 50,the filter update interval to 6,and the number of conjugate gradient iterations to 5. In this paper, the ECO is implemented based on Histogram of Oriented Gradients (HOG)features. For MDNet,17we set the input size to 107 × 107,and the Intersection over Union(IoU)thresholds used to select positive and negative samples in training videos are set to 0.7 and 0.5, respectively. When using MDNet online to track the refueling drogue, the IoU thresholds used to select positive and negative samples in the first frame are set to 0.7 and 0.3,respectively. For the DaSiamRPN, the initial input size is set to 255 × 255 and gradually increases to 767 × 767 when the target is lost. The detection score thresholds for entering and leaving failure cases are set to 0.8 and 0.95, respectively. We record the precision plot and success plot and show them in Fig. 11. Additionally, we record the operating speeds on the embedded system, as shown in Table 4 (in Table 4, all algorithms are operated on Xavier AGX).

    We also test our algorithm under different weather and lighting conditions. For different conditions, we simulate the three different weather conditions: rainy, snowy and foggy weather. When simulating the weather condition, we apply a smooth random change to each of the weather conditions.The smooth random change means that the state of each weather condition varies smoothly and randomly. For exam-

    Fig. 11 shows that our algorithm has advantages over the other tracking algorithms: the success rate plot of One Pass Evaluation (OPE) and the precision plot of OPE of our algorithm are all above other object tracking algorithms. Because our tracking algorithm tracks objects by detection,it has some prior knowledge about the tracking object,and it uses the former information such as location and class to reduce false positives. Although our algorithm does not match the SiamAPN++ on speed, it has higher precision than SiamAPN++and can run up to 20.04 FPS on the embedded device to guarantee its real-time application in the AAR task.ple,when we add the rain to the clean image,we choose to randomly increase every attribute of the rain, such as its density,and tilt angle,randomly in the interval (-ε,ε)in each frame.ε is small enough to ensure that the weather state changes smoothly. After adding the weather condition to each testing video data, we test all the algorithms on these datasets and record the results in Table 5. Among these different weather conditions, foggy weather affects the tracking precision the most. Snowy weather also has effects on the precision, but not as much as foggy weather. Rainy weather has minor effects. During the test, we separately test our algorithm with and without random Gaussian noise augmentation on the snowy and rainy weather datasets. Table 5 shows that Gaussian noise augmentation can improve our tracking algorithm when dealing with snowy or rainy weather conditions. Therefore, for snowy and rainy weather conditions, our algorithm is affected the least. For foggy weather, although the performance of our algorithm drops, it still maintains the highest performance.

    Table 3 Test results of n-fold Bernoulli probability theorem based filter.

    Fig. 11 Comparison of tracking algorithms.

    Table 4 Performance comparison of tracking algorithms.

    Table 5 Test under different weather and lighting conditions.

    For the lighting condition test, we take several different lighting condition videos at night. We separately test all the algorithms on these video datasets and record the results in Table 5. In this paper, we use gamma transfer augmentation to simulate the lighting conditions for the training dataset.Therefore,our algorithm is capable of handling different lighting conditions. The testing results are recorded in Table 5,demonstrating that the performance of all the algorithms decreases when they are tested under different lighting conditions, but our algorithm still achieves the best performance among them.

    5. Conclusions

    In this paper, we first present an object detection algorithm based on YOLOv4 with the aim to increase the reliability of the detection algorithm in situations that are intolerant to mistakes and involve small objects.To improve the ability of small object detection, we use dense blocks to replace the residual blocks to improve the feature extraction capability. We also use ACNet to enrich the scale of the receptive field,which consequently increases the accuracy of the network.Additionally,we develop an n-fold Bernoulli probability theorem based filter that can decrease false positives to increase precision when tracking an object.Based on this filter and the aforementioned object detection algorithm, we develop an adaptive object tracking algorithm for the AAR task, which proves to have high precision and strong robustness with real-time operating ability in an embedded system. The experimental results demonstrate the effectiveness of the proposed tracking algorithm.

    In this paper,we consider single-object tracking tasks.This algorithm can also be used in other single-object tracking tasks.In future work,we will focus on a multi-object tracking algorithm using an object detection algorithm that can enhance the practicality of the algorithm.

    Declaration of Competing Interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    丁香六月欧美| 91麻豆精品激情在线观看国产 | e午夜精品久久久久久久| 人成视频在线观看免费观看| 国产精品秋霞免费鲁丝片| 国产成人av激情在线播放| 亚洲欧美一区二区三区国产| 亚洲第一av免费看| 国产福利在线免费观看视频| 极品人妻少妇av视频| 高潮久久久久久久久久久不卡| 久久久久网色| 久久精品aⅴ一区二区三区四区| 免费观看av网站的网址| 日韩制服骚丝袜av| 精品久久久精品久久久| 精品人妻一区二区三区麻豆| 精品国产乱码久久久久久男人| 日本色播在线视频| 久久影院123| 国产老妇伦熟女老妇高清| 欧美日韩国产mv在线观看视频| 视频在线观看一区二区三区| 欧美人与性动交α欧美精品济南到| 日韩大片免费观看网站| av天堂久久9| 99久久99久久久精品蜜桃| 又紧又爽又黄一区二区| 亚洲图色成人| kizo精华| 国产高清不卡午夜福利| 国产一区二区三区av在线| 操美女的视频在线观看| 桃花免费在线播放| 久久午夜综合久久蜜桃| 中文字幕精品免费在线观看视频| av国产精品久久久久影院| 国产免费一区二区三区四区乱码| 欧美精品啪啪一区二区三区 | 国产成人91sexporn| 久久综合国产亚洲精品| 色播在线永久视频| 老司机亚洲免费影院| 亚洲欧美精品综合一区二区三区| 在线观看www视频免费| 天天躁夜夜躁狠狠躁躁| 日本五十路高清| 国产免费又黄又爽又色| 亚洲人成网站在线观看播放| 成人亚洲精品一区在线观看| 成人亚洲欧美一区二区av| 久久久久精品国产欧美久久久 | 69精品国产乱码久久久| 中国美女看黄片| 这个男人来自地球电影免费观看| 日韩制服骚丝袜av| 亚洲国产中文字幕在线视频| 少妇的丰满在线观看| 国产欧美亚洲国产| 老汉色av国产亚洲站长工具| 色94色欧美一区二区| 久久人妻熟女aⅴ| 亚洲五月婷婷丁香| 亚洲免费av在线视频| 成人三级做爰电影| 国产一区二区 视频在线| 色播在线永久视频| 国产精品一区二区精品视频观看| 中文字幕制服av| 无遮挡黄片免费观看| 精品视频人人做人人爽| 欧美日本中文国产一区发布| 人成视频在线观看免费观看| 我的亚洲天堂| 午夜免费成人在线视频| 国产成人精品久久二区二区91| 777米奇影视久久| 一本综合久久免费| 两个人免费观看高清视频| 一级,二级,三级黄色视频| 国产精品 欧美亚洲| 97精品久久久久久久久久精品| 久久天堂一区二区三区四区| 免费人妻精品一区二区三区视频| 久久99热这里只频精品6学生| 麻豆乱淫一区二区| 亚洲国产日韩一区二区| 国产精品亚洲av一区麻豆| 2021少妇久久久久久久久久久| 在线观看一区二区三区激情| 亚洲人成电影观看| 丝袜美足系列| 国产日韩欧美亚洲二区| 大话2 男鬼变身卡| 天堂8中文在线网| 亚洲国产成人一精品久久久| 成年美女黄网站色视频大全免费| 777米奇影视久久| 国产黄色视频一区二区在线观看| 在线天堂中文资源库| 国产成人av教育| 午夜91福利影院| 女人高潮潮喷娇喘18禁视频| 国产女主播在线喷水免费视频网站| 免费av中文字幕在线| 亚洲 国产 在线| 夫妻性生交免费视频一级片| 多毛熟女@视频| 一区二区三区激情视频| 午夜福利乱码中文字幕| 国产精品国产av在线观看| 欧美精品啪啪一区二区三区 | 欧美精品一区二区大全| 亚洲一卡2卡3卡4卡5卡精品中文| 成人国产av品久久久| 久久亚洲精品不卡| 岛国毛片在线播放| cao死你这个sao货| 国产国语露脸激情在线看| 欧美精品av麻豆av| 国产一区二区在线观看av| www.熟女人妻精品国产| 欧美日韩国产mv在线观看视频| 久久精品亚洲av国产电影网| 最新的欧美精品一区二区| 亚洲精品乱久久久久久| 亚洲 欧美一区二区三区| 欧美精品亚洲一区二区| 91字幕亚洲| 国产又色又爽无遮挡免| 国产精品av久久久久免费| 久久天躁狠狠躁夜夜2o2o | 女性被躁到高潮视频| 晚上一个人看的免费电影| 久久精品aⅴ一区二区三区四区| 中文字幕人妻丝袜一区二区| 久久人妻熟女aⅴ| 最黄视频免费看| 国产一区二区 视频在线| 男人爽女人下面视频在线观看| 亚洲av成人不卡在线观看播放网 | 亚洲免费av在线视频| 欧美日韩一级在线毛片| 久久精品人人爽人人爽视色| 丝袜喷水一区| 热99久久久久精品小说推荐| 国产有黄有色有爽视频| 国产精品 国内视频| 国产精品亚洲av一区麻豆| 美女福利国产在线| 晚上一个人看的免费电影| 亚洲第一av免费看| 五月天丁香电影| 久久久久国产精品人妻一区二区| 少妇粗大呻吟视频| 欧美变态另类bdsm刘玥| 人体艺术视频欧美日本| 中文欧美无线码| 成年女人毛片免费观看观看9 | 在线亚洲精品国产二区图片欧美| 欧美人与善性xxx| 久久免费观看电影| 午夜老司机福利片| 国产片内射在线| 日日爽夜夜爽网站| 日韩 亚洲 欧美在线| 精品人妻1区二区| 久久精品国产亚洲av涩爱| 高清黄色对白视频在线免费看| 国产精品亚洲av一区麻豆| 久久久久国产精品人妻一区二区| 黄色视频不卡| 一级毛片电影观看| 97在线人人人人妻| 电影成人av| 精品少妇久久久久久888优播| 精品亚洲成a人片在线观看| 日韩av在线免费看完整版不卡| 脱女人内裤的视频| 精品人妻一区二区三区麻豆| 两性夫妻黄色片| av国产久精品久网站免费入址| 一区福利在线观看| 国产精品秋霞免费鲁丝片| 老司机在亚洲福利影院| 人人妻人人澡人人爽人人夜夜| 日韩中文字幕视频在线看片| 国产欧美日韩一区二区三区在线| 老鸭窝网址在线观看| 亚洲精品久久成人aⅴ小说| av网站免费在线观看视频| 黄片播放在线免费| 日韩 欧美 亚洲 中文字幕| 亚洲精品久久久久久婷婷小说| 少妇猛男粗大的猛烈进出视频| 天天躁狠狠躁夜夜躁狠狠躁| 一区在线观看完整版| 超碰成人久久| 看十八女毛片水多多多| 精品第一国产精品| av欧美777| 精品熟女少妇八av免费久了| 欧美日韩国产mv在线观看视频| 久久精品熟女亚洲av麻豆精品| 国产精品一国产av| 五月天丁香电影| 老汉色av国产亚洲站长工具| 又粗又硬又长又爽又黄的视频| 婷婷成人精品国产| 人成视频在线观看免费观看| 男男h啪啪无遮挡| 国产成人啪精品午夜网站| 黄色 视频免费看| videosex国产| 在线观看免费午夜福利视频| 亚洲国产日韩一区二区| 国产精品.久久久| cao死你这个sao货| 国产精品人妻久久久影院| 最近手机中文字幕大全| 黑人巨大精品欧美一区二区蜜桃| 十八禁网站网址无遮挡| 久久精品成人免费网站| 激情五月婷婷亚洲| 天天躁夜夜躁狠狠躁躁| 最新的欧美精品一区二区| 极品人妻少妇av视频| 高清欧美精品videossex| 亚洲男人天堂网一区| 各种免费的搞黄视频| 丰满饥渴人妻一区二区三| 国产成人精品在线电影| av国产精品久久久久影院| 999精品在线视频| 天堂中文最新版在线下载| 亚洲国产中文字幕在线视频| 美国免费a级毛片| 国产真人三级小视频在线观看| av网站免费在线观看视频| 国产精品偷伦视频观看了| 悠悠久久av| 午夜免费观看性视频| 欧美精品一区二区大全| 亚洲情色 制服丝袜| 中文字幕制服av| 大话2 男鬼变身卡| 日韩大片免费观看网站| 在线观看国产h片| 侵犯人妻中文字幕一二三四区| 国产精品国产av在线观看| 国产精品国产三级专区第一集| 在线观看国产h片| 国产成人欧美| 久久亚洲国产成人精品v| 国产成人91sexporn| 热re99久久精品国产66热6| 男女免费视频国产| 最近手机中文字幕大全| av天堂在线播放| 母亲3免费完整高清在线观看| 国产免费视频播放在线视频| 亚洲国产看品久久| a级毛片在线看网站| av片东京热男人的天堂| 精品免费久久久久久久清纯 | 最新的欧美精品一区二区| 成人影院久久| 日韩 亚洲 欧美在线| 男女床上黄色一级片免费看| 18禁裸乳无遮挡动漫免费视频| 婷婷色av中文字幕| 一区二区三区精品91| www日本在线高清视频| 欧美激情极品国产一区二区三区| 9色porny在线观看| 国产精品国产av在线观看| 狂野欧美激情性xxxx| 午夜视频精品福利| 欧美精品一区二区免费开放| 亚洲一区中文字幕在线| 最近中文字幕2019免费版| 久久精品久久精品一区二区三区| 国产精品二区激情视频| 青春草亚洲视频在线观看| 免费少妇av软件| 久久这里只有精品19| 婷婷色av中文字幕| 亚洲激情五月婷婷啪啪| a 毛片基地| 一二三四社区在线视频社区8| 欧美另类一区| 人妻 亚洲 视频| 国产97色在线日韩免费| 国产在线一区二区三区精| 欧美亚洲 丝袜 人妻 在线| 久久精品亚洲熟妇少妇任你| 中文字幕人妻丝袜制服| 日韩免费高清中文字幕av| 19禁男女啪啪无遮挡网站| 欧美大码av| 精品视频人人做人人爽| 欧美日韩亚洲国产一区二区在线观看 | 在线天堂中文资源库| bbb黄色大片| 最近手机中文字幕大全| 国产主播在线观看一区二区 | 97精品久久久久久久久久精品| 中文字幕人妻丝袜一区二区| 精品久久蜜臀av无| 在线观看人妻少妇| 日韩大片免费观看网站| 丝袜在线中文字幕| 蜜桃在线观看..| 国产精品一区二区精品视频观看| 免费少妇av软件| 国产成人精品无人区| 成人18禁高潮啪啪吃奶动态图| 亚洲精品中文字幕在线视频| 19禁男女啪啪无遮挡网站| 亚洲国产av新网站| 老司机影院成人| 99精国产麻豆久久婷婷| 91精品伊人久久大香线蕉| 十八禁网站网址无遮挡| 亚洲精品久久午夜乱码| 日韩,欧美,国产一区二区三区| 中文字幕最新亚洲高清| 久久亚洲国产成人精品v| 国产精品三级大全| 精品一品国产午夜福利视频| 国产精品 欧美亚洲| 美女中出高潮动态图| 最新在线观看一区二区三区 | 色94色欧美一区二区| 国精品久久久久久国模美| 国产精品二区激情视频| xxxhd国产人妻xxx| 日本黄色日本黄色录像| 日日爽夜夜爽网站| videosex国产| 亚洲国产最新在线播放| 少妇裸体淫交视频免费看高清 | 一区二区三区激情视频| 亚洲美女黄色视频免费看| 麻豆乱淫一区二区| 亚洲美女黄色视频免费看| 亚洲三区欧美一区| 国产色视频综合| 久久久精品区二区三区| 成人黄色视频免费在线看| 另类精品久久| 婷婷色综合大香蕉| 三上悠亚av全集在线观看| 亚洲天堂av无毛| 欧美亚洲日本最大视频资源| 日本色播在线视频| 男人舔女人的私密视频| 日韩伦理黄色片| 免费在线观看影片大全网站 | 97精品久久久久久久久久精品| 欧美精品人与动牲交sv欧美| 侵犯人妻中文字幕一二三四区| 色精品久久人妻99蜜桃| 亚洲五月色婷婷综合| 免费看十八禁软件| 日韩中文字幕欧美一区二区 | 日本五十路高清| 极品人妻少妇av视频| 亚洲精品一区蜜桃| 国产成人av教育| 韩国精品一区二区三区| 久久久久久亚洲精品国产蜜桃av| 国产精品麻豆人妻色哟哟久久| 亚洲精品乱久久久久久| 日韩精品免费视频一区二区三区| 国产男人的电影天堂91| 久久久精品免费免费高清| 欧美国产精品一级二级三级| 黄片小视频在线播放| 亚洲情色 制服丝袜| 国产精品二区激情视频| 免费一级毛片在线播放高清视频 | 成年人免费黄色播放视频| 久久精品久久精品一区二区三区| 性高湖久久久久久久久免费观看| 色播在线永久视频| 成年美女黄网站色视频大全免费| 欧美国产精品va在线观看不卡| 久久99一区二区三区| 久久久久网色| 汤姆久久久久久久影院中文字幕| 热re99久久精品国产66热6| 波多野结衣一区麻豆| 欧美 日韩 精品 国产| 操出白浆在线播放| 自线自在国产av| 日本vs欧美在线观看视频| 精品亚洲成a人片在线观看| 18禁裸乳无遮挡动漫免费视频| 亚洲第一av免费看| 亚洲精品国产一区二区精华液| 精品人妻在线不人妻| svipshipincom国产片| 国产91精品成人一区二区三区 | 男女高潮啪啪啪动态图| 国产精品二区激情视频| 男男h啪啪无遮挡| 最近中文字幕2019免费版| 无限看片的www在线观看| 亚洲七黄色美女视频| 欧美日韩av久久| 国产黄色免费在线视频| 欧美日韩黄片免| 69精品国产乱码久久久| 日韩制服丝袜自拍偷拍| 首页视频小说图片口味搜索 | 高潮久久久久久久久久久不卡| 国产黄色视频一区二区在线观看| 汤姆久久久久久久影院中文字幕| 丝袜在线中文字幕| 99久久精品国产亚洲精品| 亚洲av在线观看美女高潮| 只有这里有精品99| 99国产精品99久久久久| 美国免费a级毛片| 国产一卡二卡三卡精品| 亚洲欧美成人综合另类久久久| 狂野欧美激情性bbbbbb| 久久中文字幕一级| 9色porny在线观看| 欧美97在线视频| 日韩一卡2卡3卡4卡2021年| 黄色 视频免费看| 老鸭窝网址在线观看| 亚洲欧美清纯卡通| 亚洲精品美女久久av网站| 热re99久久国产66热| 人体艺术视频欧美日本| 老司机在亚洲福利影院| 午夜av观看不卡| 午夜精品国产一区二区电影| 各种免费的搞黄视频| h视频一区二区三区| 久久影院123| 欧美激情极品国产一区二区三区| 最近手机中文字幕大全| 1024视频免费在线观看| 美女视频免费永久观看网站| 七月丁香在线播放| 午夜福利乱码中文字幕| 久久99热这里只频精品6学生| 日韩熟女老妇一区二区性免费视频| 最近手机中文字幕大全| 1024视频免费在线观看| 黄色 视频免费看| 少妇人妻久久综合中文| 欧美97在线视频| 中国国产av一级| 亚洲中文av在线| 亚洲自偷自拍图片 自拍| 亚洲国产欧美一区二区综合| 日本欧美视频一区| 丝袜喷水一区| 人人妻人人澡人人看| 建设人人有责人人尽责人人享有的| 精品免费久久久久久久清纯 | 国产黄色视频一区二区在线观看| 精品免费久久久久久久清纯 | 99re6热这里在线精品视频| 国产视频首页在线观看| 一区福利在线观看| 精品福利永久在线观看| 女人精品久久久久毛片| 97人妻天天添夜夜摸| 大型av网站在线播放| 大片免费播放器 马上看| 国产欧美日韩精品亚洲av| 国产高清videossex| 精品国产一区二区三区四区第35| 国产精品亚洲av一区麻豆| 精品人妻一区二区三区麻豆| 亚洲欧美成人综合另类久久久| 每晚都被弄得嗷嗷叫到高潮| 制服诱惑二区| 丰满饥渴人妻一区二区三| 最黄视频免费看| 午夜免费成人在线视频| 最黄视频免费看| 性少妇av在线| 欧美 日韩 精品 国产| 久久性视频一级片| 下体分泌物呈黄色| 久久人妻福利社区极品人妻图片 | 巨乳人妻的诱惑在线观看| 一区二区av电影网| 少妇粗大呻吟视频| 久久九九热精品免费| 亚洲欧美激情在线| 日韩大码丰满熟妇| 丝袜脚勾引网站| 伊人亚洲综合成人网| 亚洲国产精品成人久久小说| 黄频高清免费视频| 不卡av一区二区三区| 国产熟女欧美一区二区| 咕卡用的链子| 岛国毛片在线播放| www.熟女人妻精品国产| 国产成人系列免费观看| 精品久久久久久电影网| av天堂在线播放| 老司机午夜十八禁免费视频| 丝瓜视频免费看黄片| 亚洲欧美成人综合另类久久久| 午夜精品国产一区二区电影| 免费在线观看黄色视频的| 老司机靠b影院| 成年女人毛片免费观看观看9 | 成人三级做爰电影| 国产欧美日韩精品亚洲av| 亚洲自偷自拍图片 自拍| 国产精品偷伦视频观看了| 涩涩av久久男人的天堂| av天堂久久9| 午夜福利在线免费观看网站| 狂野欧美激情性bbbbbb| 视频区图区小说| 日韩欧美一区视频在线观看| 国产成人免费无遮挡视频| 交换朋友夫妻互换小说| 免费在线观看影片大全网站 | 午夜免费成人在线视频| 国产片内射在线| 成人免费观看视频高清| 国产精品99久久99久久久不卡| 婷婷色综合大香蕉| 日韩 欧美 亚洲 中文字幕| 脱女人内裤的视频| 国产成人精品久久二区二区91| 欧美亚洲日本最大视频资源| 亚洲国产欧美在线一区| 精品少妇黑人巨大在线播放| 我要看黄色一级片免费的| 在线观看免费日韩欧美大片| 午夜福利视频在线观看免费| 黑人欧美特级aaaaaa片| 成年美女黄网站色视频大全免费| 亚洲精品国产区一区二| 久久久久视频综合| 精品人妻熟女毛片av久久网站| 我要看黄色一级片免费的| 国产精品一国产av| 日韩视频在线欧美| 黄色毛片三级朝国网站| 亚洲欧美一区二区三区黑人| 欧美日韩黄片免| 成人三级做爰电影| 国产三级黄色录像| 国产精品久久久久久精品电影小说| 久9热在线精品视频| 国产xxxxx性猛交| 久久青草综合色| 亚洲黑人精品在线| 18禁观看日本| 亚洲精品国产色婷婷电影| 美女脱内裤让男人舔精品视频| 丰满饥渴人妻一区二区三| 成年女人毛片免费观看观看9 | 性少妇av在线| 免费少妇av软件| 一级毛片电影观看| 欧美激情 高清一区二区三区| av天堂久久9| 深夜精品福利| 99久久99久久久精品蜜桃| 国产成人精品久久久久久| 亚洲午夜精品一区,二区,三区| 中国美女看黄片| 欧美成人精品欧美一级黄| 亚洲国产av影院在线观看| 国产xxxxx性猛交| 欧美中文综合在线视频| 熟女av电影| 啦啦啦在线免费观看视频4| 国产精品.久久久| 国产免费一区二区三区四区乱码| 你懂的网址亚洲精品在线观看| 欧美日韩精品网址| 十八禁人妻一区二区| 飞空精品影院首页| 十分钟在线观看高清视频www| 国产成人系列免费观看| 一级毛片电影观看| 国产一区二区三区综合在线观看| 欧美精品啪啪一区二区三区 | 欧美成狂野欧美在线观看| 日韩制服丝袜自拍偷拍| 啦啦啦在线观看免费高清www| 亚洲精品一二三| 好男人视频免费观看在线| cao死你这个sao货| 蜜桃国产av成人99| 国产成人91sexporn| 人妻 亚洲 视频| 啦啦啦视频在线资源免费观看| 亚洲七黄色美女视频| 激情视频va一区二区三区| 精品国产一区二区久久| h视频一区二区三区| 亚洲色图综合在线观看| 免费观看人在逋| 亚洲精品成人av观看孕妇| 午夜老司机福利片| av欧美777| 啦啦啦视频在线资源免费观看|