• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Online multi-target intelligent tracking using a deep long-short term memory network

    2023-10-25 12:12:52YongqunZHANGZhenyunSHIHongingJIZhenzhenSU
    CHINESE JOURNAL OF AERONAUTICS 2023年9期

    Yongqun ZHANG, Zhenyun SHI, Honging JI, Zhenzhen SU

    a School of Electronic Engineering, Xidian University, Xi’an 710071, China

    b School of Computer Science and Technology, Xidian University, Xi’an 710071, China

    KEYWORDS

    Abstract Multi-target tracking is facing the difficulties of modeling uncertain motion and observation noise.Traditional tracking algorithms are limited by specific models and priors that may mismatch a real-world scenario.In this paper, considering the model-free purpose, we present an online Multi-Target Intelligent Tracking (MTIT) algorithm based on a Deep Long-Short Term Memory (DLSTM) network for complex tracking requirements, named the MTIT-DLSTM algorithm.Firstly, to distinguish trajectories and concatenate the tracking task in a time sequence,we define a target tuple set that is the labeled Random Finite Set (RFS).Then, prediction and update blocks based on the DLSTM network are constructed to predict and estimate the state of targets,respectively.Further,the prediction block can learn the movement trend from the historical state sequence,while the update block can capture the noise characteristic from the historical measurement sequence.Finally, a data association scheme based on Hungarian algorithm and the heuristic track management strategy are employed to assign measurements to targets and adapt births and deaths.Experimental results manifest that, compared with the existing tracking algorithms, our proposed MTIT-DLSTM algorithm can improve effectively the accuracy and robustness in estimating the state of targets appearing at random positions, and be applied to linear and nonlinear multi-target tracking scenarios.

    1.Introduction

    Multi-target tracking technology is an important branch of information fusion, and its main purpose is to continuously estimate the number of targets and their states from a sequence of observations in a scenario.1It has a wide application prospect not only in military fields2such as aerial reconnaissance,missile defense,and battlefield surveillance,but also in civilian fields such as robot vision,3traffic control, and autonomous vehicles.4Generally speaking, multi-target tracking mainly faces the following challenges: the number of targets is timevarying due to births and deaths; the accuracy of estimated states is usually degraded by measurement errors and uncertainty of target motions.In addition, data association, detection uncertainty, and false alarms can also bring obstacles.5At present, tracking multiple targets in an unconstrained scenario is still a very difficult task.

    A classic multi-target tracking framework is based on Bayesian theory, which has been widely used and achieved good tracking accuracy.Typical algorithms include Multiple Hypothesis Tracking (MHT),6Joint Probabilistic Data Association (JPDA),7Random Finite Set (RFS) filtering,8etc.At present, the most concerned multi-target tracking algorithm is RFS filtering, which recursively propagates the posterior density of target states forward.To address the heavy computational load of RFS filtering,the Probability Hypothesis Density (PHD),9Cardinalized PHD (CPHD),10Multi-target Multi-Bernoulli (MeMBer),8and Cardinality-Balanced MeMBer(CBMeMBer)11filters have been successively developed as approximations.Recently, a Poisson Multi-Bernoulli Mixture(PMBM) filter12is proposed to handle multi-target tracking with higher tracking accuracy.However,the PMBM filter suffers from the heavy computational burden due to the global association history hypotheses.In addition,to further improve the tracking performance of RFS filtering algorithms, several innovative algorithms13–16have been proposed in our early works.Strictly speaking, these tracking algorithms are not multi-target trackers because they cannot distinguish each target.To this end, Generalized Labeled Multi-Bernoulli(GLMB)17,18and LMB19filters are proposed, which formally estimate target tracks by applying the labeled RFS formulation.All of these algorithms require a given system model during initialization, including a target state transition model and a sensor observation model.The former is used to predict the target state at the next time step,while the latter is used to map the predicted state to the measurement space for correction.If the system model does not match the unknown environment,the tracker will fail to track targets.

    Recently, deep learning based on neural networks due to powerful nonlinear mapping, learning, and computing power has been growing rapidly and showing great advantages in many fields, such as image recognition,20speech processing,21video processing,22etc.However,there has been relatively less work applying deep learning to multi-target tracking.Milan et al.firstly proposed an end-to-end multi-target tracking algorithm based on deep learning.23Subsequently, Emambakhsh et al.used an Long-Short Term Memory (LSTM) network to predict the state for each target.24To deal with the combination explosion of multi-target data association, Liu et al.employed an LSTM network to learn the association weights between targets and measurements, avoiding traversing all possible association events.25Additionally, Gao et al.presented LSTM-based deep recurrent neural networks to track a single target with linear and nonlinear motions.26,27At present, multi-target tracking algorithms based on deep learning are limited to radar and pedestrian targets in ideal scenarios,and they have not performed well in more complex scenarios,such as target maneuvering, poor cooperation, etc.Consequently, figuring out how to combine the advantages of the deep learning theory and a neural network structure to reliably track multiple targets in complicated scenarios is an urgent issue to be solved.

    Inspired by the potential of recurrent neural networks, we propose a Multi-Target Intelligent Tracking(MTIT)algorithm based on a Deep LSTM (DLSTM) network with a recurrent structure to online track each labeled target, i.e., a MTITDLSTM algorithm.It is here emphasized that the proposed data-driven MTIT-DLSTM algorithm does not need to model motions and rely on priors,which is more suitable for scenario diversity.Thus, the proposed algorithm is capable of dynamically tracking multiple targets in complex scenarios involving clutter, birth, and death of targets, as well as linear and nonlinear motions.The main contributions of our work are summarized as follows:

    (1) We design a prediction block based on a DLSTM network to predict the state of each target, since the estimation of the target state for continuous time steps can be regarded as a sequential learning problem.The prediction block is required to capture the motion trend from historical states and transfer target states to the next time step.Therefore,the MTIT-DLSTM algorithm can adaptively track multiple targets moving in different ways.

    (2)We design an update block based on DLSTM to update the state of each target.The main purpose of the update network is to learn the noise characteristics from historical measurements and update states.

    (3) We build a training set corresponding to the network structure in terms of different kinds of target motions to ensure that the network can deal with tracking of multiple targets in complex scenarios.Adam optimizer is then adopted to train prediction and update networks,while a cosine annealing algorithm is used to dynamically change the learning rate to ensure convergence of the model.

    (4) To handle the time-varying number of targets, we use heuristic track management to adapt births and deaths.Meanwhile, to find the corresponding measurements for targets,data association based on Hungarian algorithm is used to seek the association map between measurements and targets.

    The paper is organized as follows.First of all,the necessary background knowledge about target tracking and DLSTM networks are reviewed in Section 2.Then,Section 3 establishes the overall algorithm framework and introduces the involved prediction and update networks, data association, track management, update survivals, and assign births in detail, respectively.Subsequently, how to train networks is explained in Section 4.Finally, Section 5 is dedicated to experimental results,and Section 6 concludes the paper,giving future working directions.

    2.Background

    A brief review of target tracking and DLSTM is given in Sections 2.1 and 2.2,respectively, which is the theoretical basis of the proposed MTIT-DLSTM algorithm.

    2.1.Target tracking

    Target tracking is the ultimate goal of radar back-end data processing, composed of point aggregation, track initiation,track association, track fusion, track extinction, and so on.Traditional target trackers based on Bayesian filtering are to recursively compute the reliability of various values, i.e., the posterior probability of target states given measurements.28In the state space and measurement space,the motion and observation equations are, respectively, modeled as

    where xkand zkdenote the target state and measurement vectors at time k, respectively; wk-1and vkrepresent the process noise and observation noise, respectively, which are independent of each other.The motion model, defined by Eq.(1),describes how a target moves, while the observation model,given by Eq.(2), describes the relationship between the target state and measurement.

    Bayesian filtering consists of two procedures: prediction to compute the initial density using a motion model; update to compute the posterior density using a measurement model,i.e.,

    where fk|k-1(xk|xk-1)is the transition probability density determined by the motion equation,while gk(zk|xk)is the likelihood function determined by the observation equation.The detailed process of Bayesian filtering is shown in Fig.1.

    It is worth noting that all information about the state at time k is encapsulated in the posterior density determining the final estimation accuracy.Prediction and update are carried out recursively to achieve a continuous estimation of the target state.In general, to alleviate the computational burden of the above equations,Kalman filter29and the particle filter30are adopted to provide two main implementation approaches.The former performs under linear and Gaussian assumptions for the state and measurement models,while the latter approximates the arbitrary distributions using sequential importance sampling.

    Under assumptions of the linear system and Gaussian noise, the state transition and measurement equations can be described as

    where Fk-1is the state transition matrix, and Hkis the transformation matrix from the state space to the observation space; wk-1and vkare zero mean Gaussian noise with covariances Qk-1and Rk, respectively.

    Considering a two-dimensional tracking scenario,the Constant Velocity(CV)and Constant Turn(CT)motions are modeled as

    where Tsis the sampling interval,and Ω is the turn rate.Thus,if a target moves at CV,its state vector is xk=[x,vx,y,vy]T;if a target moves at CT, its state vector is xk=[x,vx,y,vy,Ω]T.

    Remark 1.In order to make the proposed algorithm get rid of the limitations of the model, we utilize neural networks of deep learning to predict and update via learning motion trends and noise characteristics from the historical positions of a target,which correspond to estimating the conditional probabilities p(xk|x1:k-1) and p(xk|z1:k), respectively.Therefore, to adapt various motion modes,the state vector is defined as xk?[x,y]Tin this work.

    2.2.DLSTm

    2.2.1.Recurrent neural network

    Fig.1 Bayesian filtering process.

    As previously analyzed, the target tracking problem is essentially to estimate target states from the measurement sequence.Thus, target tracking can be realized by a Recurrent Neural Network (RNN) that can map sequential measurements to true states.A typical neural network consists of one input layer, one or several hidden layers, and one output layer.31An RNN is a kind of particular neural network with a recurrent structure, through which historical information can be stored and exploited.In the RNN structure, the hidden layer neurons at different times are connected via the hidden state h carrying the historical information,32as shown in Fig.2.For a sequence-to-sequence task, the RNN can produce a specific output sequence, given an input sequence with a temporal structure33.

    The trained RNN is capable of estimating the output sequence (y1,y2,???,yK′) given an input sequence (ε1,ε2,???,εK),on condition of maximizing the conditional density p(y1,y2,???,yK′|ε1,ε2,???,εK), where K′may differ from K.This conditional density is represented as a product of the probability densities via the chain rule,34i.e.,

    where v is the fixed joint representation of the input sequence and last hidden state of the RNN.Each p(yk|v,y1,y2,???,yk-1)represents a recurrent module defined by a deterministic function33.

    Remark 2.The above analysis provides theoretical support for estimating optimal states of targets by means of an RNN.In this work, we will mainly estimate the states that maximize the conditional densities p(xk|x1:k-1) and p(xk|z1:k) mentioned in Section 2.1, which can be implemented by a chain of recurrent modules.Obviously,the output of the last module is just what we need.

    2.2.2.DLSTM network

    A vanishing gradient phenomenon of an RNN will occur when training data with long temporal dependencies.35To avoid this phenomenon, researchers have designed some modified loop units to improve the traditional RNN, such as LSTM combined with the ‘‘gate” mechanism.36A typical LSTM unit structure is shown in Fig.3.

    The core of the LSTM unit consists of two information transmission lines (the cell state ckand the hidden state hk)and three ‘‘gates”, which are usually composed of a sigmoid neural network layer and a pointwise multiplication operation.The cell state ckand the hidden state hkupdated by

    where ⊙represents the pointwise multiplication operation;fkis the forgetting decision vector, multiplied by ck-1to determine which information from the previous time should be forgotten;ikis the input decision vector, multiplied by c~k to determine which information at the current time should be added to the cell state ck; okis the output decision vector, multiplied by tanh(ck) to determine which information in ckshould be added to the hidden state,and okare,respectively,calculated as

    Fig.2 Recurrent neural network.

    Fig.3 Structure of an LSTM network.

    where W and b are,respectively,the weight matrix and the bias vector, obtained through network training.Note that these parameters must be optimized to minimize the predictive error of the network.σ(?) and tanh(?) represent the sigmoid activation function and the hyperbolic tangent activation function,respectively, written as

    A deep model can be more efficient at representing some functions than a shallow model.37The RNN introduced earlier has only one hidden layer, while stacking more than two hidden layers will get a deep RNN.Similarly,the LSTM network can also be extended in time and space dimensions to obtain a Deep LSTM(DLSTM)network,as shown in Fig.4,which can learn more hidden information to increase the expressive power.

    Remark 3.In this work, in order to enable the network to map sequential measurements to true states,an RNN-based structure is considered, as analyzed by Eq.(9).Specifically, the DLSTM network will be applied due to its tremendous superior in extracting features from the input sequence.

    3.Multi-target intelligent tracking with DLSTM network

    In this section, we present multi-target intelligent tracking based on the DLSTM network, called the MTIT-DLSTM algorithm.Firstly, in Section 3.1, we define a problem formulation and a target tuple set in Sections 3.1.1 and 3.1.2 as the theoretical basis of the proposed MTIT-DLSTM algorithm,respectively.Then, an online intelligent tracking framework of the MTIT-DLSTM algorithm is constructed in Section 3.2.Subsequently,Section 3.3 elaborates the network construction process of the most important prediction and update blocks.Finally,data association,track management,update survivals,and assign births, involved in the MTIT-DLSTM algorithm framework, are illustrated in Sections 3.4-3.6, respectively.

    3.1.Definition of a target tuple set

    3.1.1.Problem formulation

    Fig.4 Structure of a DLSTM network.

    3.2.Online intelligent tracking framework

    Different from the framework of recursive Bayesian filtering,we here use two DLSTM networks to implement prediction and update separately.Besides, data association based on Hungarian algorithm is used to assign measurements for targets.The proposed framework has two main advantages: (A)prediction and update are two components that can be trained individually; (B) the framework becomes modular, making it easy to replace another module.The overall framework of the proposed MTIT-DLSTM algorithm is shown in Fig.5.It can be seen that multi-target intelligent tracking based on DLSTM is online, connected with the target tuple set Dk,meaning that the output of the kth iteration will be the partial input of the (k+1)th iteration.The overall process will be introduced in detail as follows.

    Firstly, the cumulative historical state setis given into the predictor, which loads the trained DLSTM model for each target and predicts the target state set Xk|1:k-1in a d-dimension space.Since the model is universal to all targets, multiple targets treated as a batch can be fed together into the predictor, which greatly speeds up the learning efficiency and can meet real-time requirements.

    Secondly, to enable the targets to match the corresponding measurements, we compute the pairwise-distance matrix Ck?RMk-1×Nkfor data association, where each element is the statistical distance between the predicted state of track ?iand measurement zk,j.The output of the data association block based on Hungarian algorithm is an assignment matrix Akwith the same size as that of Ck.In other words, we obtain an association map θk(?) equivalent to Ak.This is followed by track management to remove deaths, obtain survivals,and assign births with the assistance of the weight set, defined by Eq.(21).Then, put the labels of survivals and births into sets Ls,kand Lb,k, respectively.

    Thirdly,the update block aims to estimate the states of targets in Ls,k,while the assignment birth block is to assign target tuples for targets in Lb,k.There are two ways to update,depending on whether the measurement is assigned to the target., produced through appending zk,θk(?i)to, is fed into the update network to estimate the target state when θk(?i)>0,while the update is performed without any measurement but with the target’s predicated state when θk(?i)=0.Consequently, the union Dkof the birth and survival target tuple sets are regarded as the output for the kth iteration.

    Finally,the target state setis extracted from Dkaccording to the survival weight wmin.

    It is noteworthy that the crux of the proposed MTITDLSTM algorithm is the network construction of prediction and update blocks,which directly affects the multi-target intelligent tracking performance.Subsequently, we will give their specific construction processes in the following subsections.

    Fig.5 Overall framework of proposed MTIT-DLSTM algorithm.

    3.3.DLSTM-based prediction and update blocks

    3.3.1.Prediction block

    In Bayesian filtering, the task of prediction is to transfer the target state from time k-1 to time k by means of a motion model underlying Markov hypothesis.29If the motion model does not match the target motion, the prediction will be paralyzed,resulting in failure to track the target.DLSTM networks provide a data-driven solution to this challenge by learning motion trends intelligently from previous states.

    In this work, a DLSTM network with 10 time steps and 2 hidden layers is used to predict, which is illustrated in Fig.6.It can be seen that the prediction network retains the highorder Markov transition state.To predict the state of a mature target, the input and output of the prediction block are represented as a sequence and a vector,respectively,indicating that this block is a many-to-one structure.As a matter of fact, the DLSTM network is a many-to-many structure, as shown in Fig.4, but the prediction block, designed by us, takes only the output of the last time step as the output here.

    In general, the length of the input sequence is variable.Obviously,the motion trend is reflected in at least two continuous time steps.In the first two moments, we choose the state of the first moment as the prediction of the second moment.Besides, we hold the view that the historical information from the past 10 time steps is sufficient for learning motion trends.Therefore, the minimum and maximum lengths of the input sequence in the prediction block are set to 2 and 10, respectively.During the first few time steps of tracking,the predicted states will be slightly inaccurate due to the lack of the historical information available.With the accumulation of more and more historical information,the predicted states become closer and closer to the true states.

    As shown in Fig.6, the states of the input sequence are firstly normalized, which makes the model converge faster.Supposing that the motion scenario is a larger rectangle in the two-dimensional space, the state can be normalized by

    where rmaxand rminrepresent the boundary extremum points of the scenario, i.e., (x1,y1) and (x2,y2), as shown in Fig.7.Then, the normalized sequence is fed into the DLSTM network to extract motion characteristics.Finally, its output is processed to obtain the predicted states via a single-layer linear network.Subsequently, these states will be used to match correct measurements for targets and participate in state estimating.

    It is worth noting that the prediction network can directly handle the input from multiple targets because the input of DLSTM is provided as a tensor consisting of ‘Batch’, ‘Seq’,and ‘Feature’.For the above three parameters, ‘Batch’ and‘Seq’are variable,and‘Feature’is fixed,where‘batch’is equal to the number of targets to be tracked, ‘Seq’ is equal to the length of the historical sequence, and ‘Feature’ is the dimension of the target state.Taking two targets as example,Fig.8 shows the process of assembling the input of the prediction network.In Fig.8, the maximum value of ‘Seq’ is set to 10.When preparing the input of the prediction network, the historical information should be truncated for targets’ surviving length being greater than 10 time steps,while the historical information should be filled with initial states for targets’ surviving length being less than 10 time steps.

    3.3.2.Update block

    The noise,attached to measurements from a trajectory,is generally supposed to follow a Gaussian distribution, whose covariance is influenced by the sensor and environment.Traditional tracking filters rely on priors about noise to estimate the target state.However, in complex tracking environments, a large number of hypotheses and priors cannot be maintained.In this work, in order to solve this problem, we design an update network based on DLSTM to recover states from polluted measurements by capturing noise characteristics from historical measurements, as shown in Fig.9.

    Fig.6 Prediction network.

    Fig.7 Boundary of the two-dimensional scenario.

    The update network,depicted by Fig.9,has a similar structure to that of the prediction network.The only difference is that the number of time steps is more than that of the prediction network.Since the statistical properties of noise are hidden in enough samples, the DLSTM network needs to learn noise information from a longer historical measurement sequence.The length of the input sequence is variable,ranging from 10 to 20.In this work, in order to ensure tracking accuracy,Kalman filter is used to estimate the state when the length of the cumulative measurement sequence is less than 10.Its detailed process is summarized as follows:

    where Pk|k-1and Pkare, respectively, the covariances of the predicted and estimated states at time k;the innovation vector ekrepresents the difference between the measurement zkand the prediction obtained by mapping the predicted state xk|k-1to the observation space; accordingly, the matrix Skis the covariance of the innovation vector ek,and Kkis Kalman gain.Substituting Eqs.(24)-(27) into Eq.(28), the equation of updating states can be obtained by

    In this work, we define x?[x,y]T, as mentioned in Remark 1,and z=[x,y]T,so Hkis the identity matrix I.Thus,Eq.(30)is transformed to

    It can be seen that when estimating the state,a larger Pk|k-1means a higher proportion of measurement, while a larger Rkmeans a higher proportion of prediction.Assuming that Pk|k-1=diag(a,a) and Rk=diag(b,b), Eq.(31) can be approximated as

    Remark 4.As mentioned earlier, the prediction block based on the DLSTM network can only learn the dynamics characteristics from the historical states, but cannot obtain the prediction covariance.Through repeatedly adjusting the coefficients and experimental verification, fx=fz=0?5 will acquire satisfactory results.Note that using the update network to estimate the target state requires historical information of at least 10 moments.In the first 10 moments,our main purpose is to ensure a continuous trajectory.During this period, the target state cannot be accurately estimated in general.Due to the lack of the priori of motion and measurement noise, classic estimating methods,such as Kalman filtering, cannot be directly used to accurately estimate the target state.In this work, in order to overcome the above shortcomings, we use fxand fzto represent the reliability of prediction and measurement, respectively.Since we cannot estimate the covariance, we have no alternative but to believe that prediction and measurement are equally reliable.

    Fig.8 Process of assembling the input of the prediction network.

    Fig.9 Update network.

    3.4.Data association

    Data association that follows the prediction block aims to assign the corresponding measurements to targets.In cluttered environment, there are ambiguities in the association due to measurements arising from true targets or false alarms.Assigning wrong measurements to targets often results in lost tracks.Moreover, clutter can produce false tracks that may rob measurements of true targets.Normally, data association schemes are used to deal with the above problems,which have received much attention in multi-target tracking research.38The purpose of data association is to compute the assignment matrix A by minimizing the global association cost, given the cost matrix C ?RM×N, which is described as25

    Subject to (s.t.)

    where i=1,2,???,M,j=1,2,???,N.The first constraint, given by Eq.(35), implies that each measurement can be assigned to at most one target, while the second one, introduced by Eq.(36), implies that a track can generate at most one measurement at any time step.Therefore, data association is a complex combinatorial optimization problem that can be handled by Hungarian algorithm.

    Usually,the target number M will be less than the measurement number N,so Hungarian algorithm will return M association maps, which means that there are all-zero columns and no all-zero rows in the above mentioned assignment matrix A.Note that an all-zero column means that the measurement does not match any existing track.In order to model missed detection,we decide to add an additional column to the assignment mapping matrix, represented as A ?RM×(N+1).Accordingly, the cost matrix C also needs to add an additional column, filled with the association threshold cthr.Hence, the optimization objective and constraint are rewritten as

    where i=1,2,???,M,j=1,2,???,N+1.

    At time k, the statistical distance is established with the weighted norm of the innovation vector39as follows:

    where ckfollows a central chi-squared distribution with n degrees of freedom,and n is the dimension of the measurement vector.Thus,the statistical distance between each survival target and each measurement is calculated to form a cost matrix.

    Using the cost matrix Ckat time k as the input,Hungarian algorithm can output the assignment matrix Ak.Then,remove the last column of Akto get the final assignment matrix.An all-zero row i means that track ?iis possibly undetected or dead, while an all-zero column j means that the jthmeasurement may be a false alarm or arise from a birth target.

    3.5.Track management

    To adapt the time-varying number of targets, it is indispensable to provide a mechanism to allow birth targets to enter the scenario and remove existing ones that disappear indefinitely.Each unassigned measurement can potentially be either the start of a new trajectory or a false alarm.Additionally,the case of a certain target with missing measurement could mean that the target has disappeared,or that the detector has failed.To address this challenge,a heuristic method,considering consecutive measurements,is proposed to adapt to the appearance and extinction of targets without priors about appearing positions.

    In general, a track goes through several stages from its appearance to extinction: track initiation, track maintenance,and track disappearance.In multi-target tracking, a point image is formed at every radar scan.If measurement points,originated from one potential birth target, appear continuously no less than 3 scanning cycles,the track formed by these points is considered as an initial track.Continuous arrival measurements maintain a mature track.When there is no new correlation measurement to maintain one track, it will enter the stage of track disappearance.Further, when one track disappears for two scan cycles, it is considered to have died.

    Track management is mainly carried out by changing the weight of each track.In this work, targets with a survival weight wk,?i= 1 are regarded as mature targets, while those at the stages of track initiation and track disappearance(0

    In Fig.10, the circle represents the survival weight that is initialized as 0.4.Note that wk,?i≥0?8 means that a new track has successfully started, marked with orange circles.Meanwhile, the letter ‘y’ refers to the situation that targets have assigned measurements, and accordingly, the survival weights will increase.The letter ‘n’ refers to the situation that targets have no assigned measurements, and accordingly, the survival weights will decrease.As we can see, it is strict for raw and dying targets to reduce their weights to 0, which helps to prevent false tracks.However, it is tolerant for mature targets to reduce their weights to 0.8 instead of 0, which aims at dealing with missed detection.Besides, track management needs to append the labels of targets with a survival weight greater than 0 (wk,?i>0) to the set Ls,k.Meanwhile, unassigned measurements, possibly originated from birth targets or false alarms,are temporarily treated as raw targets with an initial survival weight winit=0?4.At last, assign the labels for raw targets and append them to the set Lb,k.For completeness,we summarize the key steps of track management in Algorithm 1.

    Algorithm 1.Pseudocode for track management.

    Input: Ak,winit,{wk-1,?i}Mk-1 i=1 Output: Ls,k, Lb,k 1.Initialize: Ls,k =Lb,k ={}; n=0; convert Ak to association map θk(?);2.for i=1,2,???,Mk-1 do 3.if wk-1,?i =1 then /*mature targets */4.append ?i to Ls,k;5.if θk(?i)>0 then 6.wk,?i ←1 7.else 8.decrease survival weight;9.end 10.end 11.if 00 then 13.increase survival weight, and append ?i to Ls,k;14.else 15.wk,?i ←0 16.end 17.end 18.end 19.for j=1,2,???,Nk do 20.if Σ iaij =0 then /* possible birth target or clutter*/21.n ←n+1; ?n ←(k,n); wk,?n ←winit;22.append ?n to Lb,k;23.end 24.end 25.return

    3.6.Update survivals and assign births

    Fig.10 Dynamic change of survival weight in track management.

    Algorithm 2.Pseudocode for update survivals.

    Input: Ls,k,{S(?i)z,k-1}Mk-1 i=1 Output: Ds,k 1.for ? in Ls,k do 2.if θk(?)>0 then 3.append zk,θk(?) to S(?)z,k-1, resulting in S(?)z,k;4.if len(S(?)z,k)>10 then 5.^xk,? ←update net(S(?)z,k) /*use update network to update */6.else 7.^xk,? ←fx×xk|1:k-1,?+fz×zk,θk(?)8.end 9.end 10.if θk(?)=0 then /* regard prediction as optimal estimation */11.^xk,? ←xk|1:k-1,?12.end 13.append ^xk,? to S(?)x,k-1, resulting in S(?)x,k;14.ds,k ←(S(?)x,k,S(?)z,k,wk,?)?Ds,k 15.end 16.return

    In parallel with the above procedure, for each target from Lb,k, a target tuple is assigned as db,k=(zk,θk(?),zk,θk(?),winit)?Db,k.Therefore, the target tuple set at the kthtime step is the union of births and survivals,i.e., Dk=Ds,k∪Db,k.Finally, the target state set X^kis extracted from Dk.Set the minimum weight wmin=0?8 and then select those targets whose weights are greater than wminas estimated targets.

    4.Training of prediction and update networks

    The main innovation of our work is to use two DLSTM networks to predict and update independently.The former learns the movement trend from the historical state sequence to estimate the target state at the next moment, which is related to the accuracy of data association.The latter learns the noise characteristics from the historical measurement sequence to estimate the real state of a target, which determines the final tracking accuracy.In order to make the network have the desired function, offline pre-training is an important part.Finding optimal parameters for deep architectures still remains a harsh challenge.In this section,we will point out some of the most important parameters and training details.

    4.1.Generation of training data

    Deep learning methods, owning appropriate network structures predesigned, need to extract characteristic information from massive training data.Different kinds of target motions of interest need to be sufficiently included to ensure that the network can deal with tracking of targets with different motions.In this work, we mainly consider CV and CT targets over the surveillance area [-2000,2000] m×[-2000,2000] m.The generation rules of states and measurements of targets are shown in Table 1.

    Following the above rules and dynamic, we generate 256 tracks and add Gaussian noise with different covariances to each track to generate measurements.Then, the training data is encapsulated as the input–output pairs with a time order,i.e.,

    where {Ip,k,Op,k} and {Iu,k,Ou,k} are the pair-wise sequence data used for training the prediction and update networks,respectively.Note thatoriginated from the ithtrack refers to the measurements during the previous 20 moments up to time k.It can be seen that the structure of the training data is consistent with the network structure shown in Figs.6 and 9.Further,Eqs.(40)and(41)can be understood as dividing a long track into several short tracks.Hence,the prediction and update networks are trained by 23040 (90×256) and 20736 (81×256) tracks, respectively, which is enough for the tracking requirement of this work.

    Table 1 Training data generation rules.

    4.2.Adam optimizer for training

    The ultimate goal of training the neural network is to optimize its parameters to minimize the network’s loss function.We expect the predicted and updated states to be as close to the true states as possible.Therefore, the Mean-Squared Error(MSE)is here chosen as the loss function to measure the difference between the truth and output sequences.A lower MSE implies a better prediction.Additionally, to speed up training,the training data set is divided into several batches randomly,and the parameters are updated based on the partial evaluation of the loss function in each batch.

    The trained DLSTM network can be modelled by a deterministic function FΘ(?), where Θ={θ1,θ2,???,θM}, θiis the trainable parameter shared by all LSTM units, and M is the total number of applied LSTM units.The prediction process of networks can be described as ^O=FΘ(I).Let Oiand ^Oibe the truth and estimated state sets of the ithtrack in a batch,respectively.The training objective function is defined as

    where N is the number of targets contained in a batch.Oiand Iiare known and offered by training data, given by Eqs.(40) and(41).Therefore,the loss is a function of the parameter set Θ.The purpose of the training network is to adjust parameters for many times to minimize the loss function.As mentioned in Section 3.3.1, the DLSTM network has a many-to-many structure,while the prediction block has a many-to-one structure.The loss function, as shown in Eq.(42), contains the MSE between the output of all time steps and the ground truth.We intend to make the output of each time step as close to the true value as possible by minimizing the overall loss.This can be understood as training multiple many-to-one networks at the same time.

    To obtain the neural network’s optimal parameter set Θ,Adam optimizer is adopted in this work, which is a stochastic gradient-based optimization algorithm.It differs from the traditional Stochastic Gradient Descent (SGD) algorithm in that it assigns adaptive individual learning rates to each neural network parameter separately and updates them based on the estimates of the first and second moments of the gradients.31In training, a learning rate is the step of adjusting parameters,which determines how quickly the parameters change.Note that the disadvantages of using a fixed learning rate are that a too small one will lead to slow convergence and local optimum, while a too large one will lead to failure to converge.However, when getting closer to the global minimum of the loss function,we expect that the learning rate becomes smaller to make a model parameter as close to the optimum as possible.Therefore, to train the network more effectively, we specify the initial learning rate lr=0?0005 for Adam optimizer and use the cosine annealing algorithm to dynamically change it.

    In a training epoch, all pairs of training sequences are utilized to train the networks.In order to optimize the model parameters, we intend to train the prediction and update networks for 200 epochs, which is enough to reach convergence.The changing curve of the learning rate is illustrated in Fig.11 using the cosine annealing algorithm.It can be seen that in the later stage of training, the learning rate changes slowly and gradually approaches 0, which can help to reach the minimum loss location.

    4.3.Training of networks

    In the network construction of the proposed MTIT-DLSTM algorithm,the DLSTM network for state prediction is trained with two layers and 64 hidden units.Compared to the prediction,the state estimation is a more formidable work,requiring the update network to have more representation power.To that end, the DLSTM network employed to update consists of two layers and 128 hidden units.In this work, all experiments are implemented in Python3.8 with the help of PyTorch,40using an NVIDIA GeForce RTX 3090 GPU to accelerate.The losses of the two networks during training are shown in Figs.12(a) and (b), respectively.It can be seen that the losses of the prediction and update networks tend to be stable with slight fluctuation in the later epochs, meaning that the networks have converged.

    Note that the prediction and update functions of the networks are derived from the training data.As long as the motions of a target are included in the training set,the prediction network can maintain the movement trend.Besides, as long as the measurement noise obeys the Gaussian distribution with its covariance in a consistent range with the training set,the update network can denoise from measurements.

    5.Simulation results

    In this section, in order to verify the tracking performance of the proposed MTIT-DLSTM algorithm, two complex multitarget tracking scenarios, i.e., multiple CV and CT target tracking, are considered, which involve clutter, birth, and death of targets as well as linear and non-linear motions.In addition, we compare our algorithm with popular RFSbased filters, such as the PHD and LMB filters, whose MATLAB code is available from https://ba-tuong.vo-au.com/codes.html.In the following Sections 5.2 and 5.3, some experimental results are illustrated and analyzed.

    5.1.Parameter setting and evaluation metric

    Fig.11 Learning rate versus epoch.

    Fig.12 Losses of the two networks in training.

    In all simulations,the survival probability pSand the detection probability pDare set to 0.99 and 0.98, respectively.A sensor returns position measurements, and the standard deviation of the measurement noise is set to σx=σy=10 m.The tracking period and sampling interval are set to T=100 s and Ts=1 s,respectively.The clutter measurements are uniformly distributed over the observation space and Poisson RFS with the intensity denoted by

    where u(?) denotes a uniform density over the observation region, V is calculated by V=∫u(z)dz, and λcis the mean value of Poisson distribution and set to λc=10 per scan in simulation.

    In this work, the OSPA distance is used as the metric to evaluate the performance in multi-target tracking.Considering two arbitrary sets X={x1,x2,???,xm} and Y={y1,y2,???,yn},where m,n ?N, the OSPA distance with order 1 ≤p<∞and cut-off c>0 between them is defined by41.

    where Πnis the set of permutations on {1,2,???,n}, d(c)(x,y)represents the distance between x and y,cut off c,which is calculated as d(c)(x,y)=min(c,d(x,y)).In the simulation, set p=1 and c=100.Further,the OSPA distance can be divided into a location error and a cardinality error, i.e.,

    5.2.Multiple CV target tracking

    Considering a two-dimensional surveillance region[-1000,1000] m×[-1000,1000] m, all targets with random birth and death times travel in straight lines and with different but constant velocities,as shown in Table 2.Fig.13 shows the true trajectories represented by black lines and estimated trajectories represented by other colorful lines for a single run..Different trajectories are distinguished by different colors.From Fig.13, we observe that the proposed MTIT-DLSTM algorithm can successfully track all of the targets with no track label switching, leading to an excellent single run result.However,track label switching occurs occasionally due to occlusion or false alarms, resulting in a trajectory with several colors.Note that the proposed algorithm can re-initialize a lost track at any position to ensure the continuity of tracking.

    Results are shown in Fig.14, where Figs.14(a)–(d) exhibit the cardinality estimation, OSPA distance, OSPA cardinality error, and OSPA location error of all algorithms for multiple CV target tracking, averaged over 100 Monte Carlo (MC)runs,respectively.According to Figs.14(a)and(c),in the part of estimating the number of targets, it is clear that the performance of our MTIT-DLSTM algorithm is close to that of the LMB filter and obviously better than that of the PHD filter except for the moments of birth and death.Additionally, the MTIT-DLSTM algorithm shows a larger cardinality error near targets appearing and dying, since adaptive birth and death require several consecutive moments to judge.However,the PHD and LMB filters already use the first measurement to update the static birth distributions relying on the priori.Note that the birth positions of targets in the simulation are fixed and known for the compared algorithms, while our algorithm can track targets appearing at arbitrary positions without any priori about targets and environment.

    From Fig.14(b), it can be seen that our algorithm outperforms the other algorithms in terms of the overall OSPA distance.This is due to a significant drop in the OSPA location error, as shown in Fig.14(d), attributed to the use of the prediction and update networks.The prediction network replaces the motion model to transfer the target state,while the updatenetwork replaces Kalman filter to estimate the true state,which determines the final OSPA location error.In algorithms(e.g., PHD and LMB) based on Kalman filter, the covariance reflects the reliability of the estimated state, and its convergence speed is affected by the model and noise.Different from those algorithms, the update network focuses on data and intends to remove noise from measurements to restore the real state, which clearly brings better tracking results.During the early times (0

    Table 2 Notations of important variables.

    Fig.13 Ground truth and estimated trajectories of CV targets.

    From Figs.14(c)and(d), we observe that the OSPA cardinality error increases and the corresponding location error decreases near times when a target is born(k=50 s)and a target has died(k=90 s).This phenomenon can be explained by Eqs.(45)and(46),and Eq.(45)can be further transformed towhere cost is the minimum Euclidean distance of X and Y.When targets appear, the OSPA location error drop is caused by an increase of max(m,n);when targets die,the OSPA location error drop is due to a decrease of cost.

    Table 3 Runtime comparison between all algorithms.

    Finally, the average single runtimes of all algorithms are shown in Table 3.The runtime of our MTIT-DLSTM algorithm is equivalent to that of the PHD filter, and faster than that of the LMB filter.The result implies that the proposed MTIT-DLSTM algorithm can satisfy the requirement for real-time processing in multi-target tracking.

    5.3.Multiple CT target tracking

    In this section, to verify that the proposed MTIT-DLSTM algorithm is able to handle motion uncertainty with the same trained networks, tracking multiple targets moving in a Constant Turn (CT) rate is simulated.In this scenario, there are 5 targets appearing in a two-dimensional surveillance region[-2000,2000] m×[0,2000] m with various birth and death times,as shown in Table 4.It’s worth noting that the proposed algorithm is free of model and data-driven, so there is no difference between linear and nonlinear scenarios.As mentioned in Section 2.1,both state and measurement vectors are defined as[x,y]T,which means that the proposed algorithm focuses on location information.Hence, no additional processing is required for the nonlinear scenario in this section.

    Fig.14 Simulation results for multiple CV target tracking.

    Table 4 Notations of important variables.

    Fig.15 Ground truth and estimated trajectories of CT targets.

    Results are shown in Figs.15-16,where Fig.15 exhibits the true trajectories represented by black lines and estimated trajectories represented by other colorful lines for a single run;Figs.16(a)-(d) shows the cardinality estimation, OSPA distance, OSPA cardinality error, and OSPA location error of all algorithms for multiple CT target tracking, averaged over 100 MC runs,respectively.The results are broadly in line with those of multiple CT target tracking in Section 5.2.The fluctuation of the OSPA distance results from the cardinality error.The peaks in Figs.16(b) and (c) represent the emergence or extinction of targets.As long as the number of targets remains constant, the proposed MTIT-DLSTM algorithm performs perfectly, referring to that the OSPA cardinality error is close to 0 and the OSPA location error is lower.It implies that our MTIT-DLSTM algorithm has better application prospects.

    Fig.17 Ground truth and estimated trajectory of a maneuvering target.

    5.4.Single maneuvering target tracking

    In this section, to verify that the proposed MTIT-DLSTM algorithm is able to track a maneuvering target, we design a simulation scenario including a target that moves straight at first and then turns at k=40 s.

    Fig.16 Simulation results for multiple CT target tracking.

    Fig.18 Simulation results for single maneuvering target tracking.

    Results are shown in Figs.17-18,where Fig.17 exhibits the true trajectory represented by the black line and estimated trajectory represented by blue lines for a single run, and Figs.18(a)-(d) shows the cardinality estimation, OSPA distance, OSPA cardinality error, and OSPA location error for single maneuvering target tracking, averaged over 100 MC runs,respectively.The dotted line in Fig.17 indicates the process of track initiation.From Figs.18(a)-(d), we can conclude that the proposed algorithm can accurately estimate the number of targets due to track management strategy and target states.Note that Section 3.3 mentioned that the input sequence length of the prediction network ranges from 2 to 10.Therefore, the prediction network uses historical information from the past 10 time steps at most to predict.In this section, when the target undergoes the maneuver, the prediction network also only adopts historical information from the past 10 time steps instead of especially using long history information in prediction due to the target maneuver.

    In the above three simulation experiments, codes using the MTIT-DLSTM algorithm to track are the same except for measurement data, which illustrates that the prediction network can flexibly adapt to various target motions.There are two advantages of using the prediction network for state transition:one is to replace the motion model and learn the motion trend from historical states; the other is that the prediction state of the next time is not uniquely determined by the previous time,but jointly determined by historical states,which can successfully deal with missed detection and wrong association at a certain time.

    6.Conclusions

    In our work, an intelligent tracking algorithm based on two DLSTM networks, called the MTIT-DLSTM algorithm, is proposed to address fixed motion and measurement models’problem in multi-target tracking,.We adopt LSTM-based prediction and update networks with deep structures to estimate target states from noisy measurements in a complex environment.Experimental results prove that our algorithm is able to estimate the time-varying number of targets and their tracks from measurement sets in the presences of data association uncertainty, detection uncertainty, noise, and false alarms.Moreover, in terms of robustness and accuracy, our proposed algorithm outperforms the popular PHD and LMB filters on behalf of classical algorithms.The following conclusions can be extracted from this work:

    (1) Deep learning provides a novel solution to multi-target tracking and gets rid of the dependence on the model and priors.Combined with the advantages of the DLSTM network,we predict the state at the next time based on the historical state sequence and measurement sequence, which can be tailored to scenario complexity.

    (2) Applying Adam optimization and a cosine annealing algorithm to efficiently minimize the loss makes the parameters of the model optimal.

    (3) Hungarian algorithm and heuristic track management are used to assign measurements to targets and adapt births and deaths.

    (4) The limitation of this work is that the update network only works in a rectangular coordinate system.This is due to the fact that the noise converted from a polar coordinate to a rectangular coordinate will no longer follow the Gaussian distribution.This is worthy of future research.

    Declaration of Competing Interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgements

    This work was supported by the National Natural Science Foundation of China (No.62276204), Open Foundation of Science and Technology on Electronic Information Control Laboratory, Natural Science Basic Research Program of Shanxi, China (Nos.2022JM-340 and 2023-JC-QN-0710),and China Postdoctoral Science Foundation (Nos.2020T130494 and 2018M633470).

    999久久久精品免费观看国产| 精品久久蜜臀av无| 亚洲精品美女久久久久99蜜臀| 一个人免费看片子| 超碰97精品在线观看| 麻豆乱淫一区二区| 精品少妇一区二区三区视频日本电影| 欧美人与性动交α欧美软件| 精品卡一卡二卡四卡免费| 黑人欧美特级aaaaaa片| 最近最新中文字幕大全电影3 | 午夜91福利影院| 啦啦啦在线免费观看视频4| 国产一区二区 视频在线| 一级毛片精品| 黄色毛片三级朝国网站| 亚洲国产欧美网| 色播在线永久视频| 成人精品一区二区免费| 国产成人精品久久二区二区免费| 精品国产乱子伦一区二区三区| 久久国产精品人妻蜜桃| 国产片内射在线| 国产精品欧美亚洲77777| 丰满迷人的少妇在线观看| 免费在线观看日本一区| 亚洲欧美日韩高清在线视频 | av欧美777| 80岁老熟妇乱子伦牲交| 久久精品熟女亚洲av麻豆精品| 久久国产精品男人的天堂亚洲| 日韩视频在线欧美| 亚洲美女黄片视频| 伊人久久大香线蕉亚洲五| 在线亚洲精品国产二区图片欧美| 亚洲国产欧美日韩在线播放| 久久久久久久大尺度免费视频| av天堂在线播放| 极品少妇高潮喷水抽搐| 国产真人三级小视频在线观看| 久久亚洲精品不卡| 一进一出好大好爽视频| 日韩三级视频一区二区三区| 90打野战视频偷拍视频| 国产一区二区三区在线臀色熟女 | 日本撒尿小便嘘嘘汇集6| 性少妇av在线| 成人黄色视频免费在线看| tocl精华| 成年人黄色毛片网站| 久久久久国内视频| 亚洲天堂av无毛| 一区在线观看完整版| 免费人妻精品一区二区三区视频| 母亲3免费完整高清在线观看| 最新的欧美精品一区二区| 满18在线观看网站| 久久人妻熟女aⅴ| 久久精品成人免费网站| 18在线观看网站| 免费在线观看黄色视频的| 国产成人欧美在线观看 | av国产精品久久久久影院| 王馨瑶露胸无遮挡在线观看| netflix在线观看网站| 亚洲欧美精品综合一区二区三区| 欧美+亚洲+日韩+国产| 欧美av亚洲av综合av国产av| 久久中文看片网| 国产三级黄色录像| 亚洲av日韩在线播放| 一区二区三区精品91| 极品人妻少妇av视频| 精品午夜福利视频在线观看一区 | 下体分泌物呈黄色| 在线观看www视频免费| 色在线成人网| 一区福利在线观看| 99re6热这里在线精品视频| 亚洲av电影在线进入| 亚洲精华国产精华精| 午夜精品久久久久久毛片777| 欧美人与性动交α欧美精品济南到| 视频区图区小说| www.精华液| 国精品久久久久久国模美| 欧美激情久久久久久爽电影 | 成人国产一区最新在线观看| 久久精品国产a三级三级三级| 亚洲欧美日韩另类电影网站| 精品久久久精品久久久| 精品福利观看| 丝袜在线中文字幕| 巨乳人妻的诱惑在线观看| 久久毛片免费看一区二区三区| 精品人妻熟女毛片av久久网站| kizo精华| 精品免费久久久久久久清纯 | 中文字幕精品免费在线观看视频| 欧美性长视频在线观看| 777久久人妻少妇嫩草av网站| 欧美精品亚洲一区二区| 亚洲欧美精品综合一区二区三区| 飞空精品影院首页| 亚洲全国av大片| 精品一品国产午夜福利视频| 亚洲成av片中文字幕在线观看| 日韩精品免费视频一区二区三区| 久久毛片免费看一区二区三区| 国产91精品成人一区二区三区 | 桃红色精品国产亚洲av| 中文欧美无线码| 国产一区二区在线观看av| 欧美另类亚洲清纯唯美| 淫妇啪啪啪对白视频| 99国产精品一区二区三区| 亚洲五月色婷婷综合| 亚洲精品国产区一区二| 亚洲国产毛片av蜜桃av| 日本黄色日本黄色录像| 99国产精品一区二区蜜桃av | 狂野欧美激情性xxxx| 中文字幕最新亚洲高清| 欧美日韩国产mv在线观看视频| 亚洲精品中文字幕在线视频| 三上悠亚av全集在线观看| av视频免费观看在线观看| 精品国产一区二区三区久久久樱花| 久久午夜亚洲精品久久| av又黄又爽大尺度在线免费看| 久久99热这里只频精品6学生| 色老头精品视频在线观看| 成人国产一区最新在线观看| 国产主播在线观看一区二区| 99国产精品免费福利视频| 国产男女超爽视频在线观看| 中国美女看黄片| 亚洲国产av新网站| 丰满人妻熟妇乱又伦精品不卡| 在线观看免费高清a一片| 不卡av一区二区三区| 亚洲精品成人av观看孕妇| a级毛片在线看网站| av在线播放免费不卡| 成人免费观看视频高清| 五月天丁香电影| 国产成人欧美在线观看 | 精品国产一区二区久久| 久久久国产一区二区| 精品少妇久久久久久888优播| 757午夜福利合集在线观看| 国产成人欧美| avwww免费| 国产三级黄色录像| 久久久久国内视频| 久久这里只有精品19| 国产人伦9x9x在线观看| 美国免费a级毛片| 另类精品久久| 国产真人三级小视频在线观看| 亚洲熟妇熟女久久| 大型av网站在线播放| 99热国产这里只有精品6| 精品午夜福利视频在线观看一区 | 久久久久久久大尺度免费视频| 男女无遮挡免费网站观看| 十八禁网站网址无遮挡| 久久免费观看电影| 免费不卡黄色视频| 人人澡人人妻人| 免费日韩欧美在线观看| 亚洲精品国产区一区二| 国产日韩欧美在线精品| 19禁男女啪啪无遮挡网站| 久久天躁狠狠躁夜夜2o2o| 亚洲国产av影院在线观看| 欧美精品av麻豆av| 不卡av一区二区三区| 久久久久久久大尺度免费视频| 久热这里只有精品99| 国产精品自产拍在线观看55亚洲 | 高清毛片免费观看视频网站 | 国产精品熟女久久久久浪| 亚洲五月色婷婷综合| 中文亚洲av片在线观看爽 | 老鸭窝网址在线观看| 国产极品粉嫩免费观看在线| 制服人妻中文乱码| 热99国产精品久久久久久7| 999久久久国产精品视频| 国产伦理片在线播放av一区| 亚洲欧美一区二区三区黑人| 精品国产乱码久久久久久小说| av天堂在线播放| 午夜福利乱码中文字幕| 午夜福利在线免费观看网站| 又大又爽又粗| 中文字幕色久视频| 精品国内亚洲2022精品成人 | 高清在线国产一区| 超碰成人久久| 女人高潮潮喷娇喘18禁视频| 欧美黑人精品巨大| 老司机影院毛片| 成人黄色视频免费在线看| 亚洲精品国产色婷婷电影| 99九九在线精品视频| 妹子高潮喷水视频| 人妻 亚洲 视频| 狠狠婷婷综合久久久久久88av| 精品国产乱码久久久久久小说| 欧美大码av| 欧美人与性动交α欧美软件| 韩国精品一区二区三区| 99国产精品免费福利视频| 亚洲欧美一区二区三区黑人| 别揉我奶头~嗯~啊~动态视频| 久久久久精品国产欧美久久久| 不卡av一区二区三区| 亚洲少妇的诱惑av| 国产精品免费一区二区三区在线 | 人妻 亚洲 视频| 91av网站免费观看| 午夜老司机福利片| 99精品在免费线老司机午夜| 日韩中文字幕欧美一区二区| 精品人妻在线不人妻| 国产亚洲精品久久久久5区| 香蕉国产在线看| 国产成人啪精品午夜网站| 亚洲自偷自拍图片 自拍| 午夜免费成人在线视频| 人妻一区二区av| 久久中文字幕一级| 国产欧美日韩一区二区精品| 正在播放国产对白刺激| 一区二区三区乱码不卡18| 纯流量卡能插随身wifi吗| 午夜福利乱码中文字幕| 国产免费福利视频在线观看| 久久久精品国产亚洲av高清涩受| 久久影院123| 国产精品久久久久成人av| 自线自在国产av| 精品少妇黑人巨大在线播放| 久久午夜亚洲精品久久| 成人影院久久| av不卡在线播放| 国产亚洲精品第一综合不卡| 黄色视频不卡| 欧美日韩视频精品一区| 免费久久久久久久精品成人欧美视频| 无限看片的www在线观看| 久久人人97超碰香蕉20202| 他把我摸到了高潮在线观看 | 午夜91福利影院| 日韩制服丝袜自拍偷拍| 99国产精品免费福利视频| 欧美成狂野欧美在线观看| 制服人妻中文乱码| 国产视频一区二区在线看| 国产不卡av网站在线观看| 久久免费观看电影| 欧美另类亚洲清纯唯美| 亚洲自偷自拍图片 自拍| 自线自在国产av| 午夜福利在线免费观看网站| 欧美日韩一级在线毛片| 亚洲 欧美一区二区三区| 在线观看www视频免费| 久久中文看片网| 母亲3免费完整高清在线观看| 女人精品久久久久毛片| 精品少妇内射三级| 女人被躁到高潮嗷嗷叫费观| 一进一出好大好爽视频| 色精品久久人妻99蜜桃| 欧美另类亚洲清纯唯美| 两个人看的免费小视频| 另类精品久久| 免费在线观看完整版高清| 久久99一区二区三区| 日韩中文字幕视频在线看片| 我要看黄色一级片免费的| 无遮挡黄片免费观看| 免费女性裸体啪啪无遮挡网站| 日韩中文字幕视频在线看片| www日本在线高清视频| 国产真人三级小视频在线观看| 亚洲一码二码三码区别大吗| 亚洲综合色网址| 五月天丁香电影| 在线永久观看黄色视频| 超色免费av| 国产成人欧美在线观看 | 久久人妻av系列| 亚洲国产成人一精品久久久| 大香蕉久久网| 三级毛片av免费| 香蕉久久夜色| 久久狼人影院| 国产精品成人在线| 国产免费视频播放在线视频| 热re99久久精品国产66热6| 久久av网站| 日本av免费视频播放| 久久 成人 亚洲| av线在线观看网站| 亚洲九九香蕉| 91麻豆av在线| 黄色视频在线播放观看不卡| 热99久久久久精品小说推荐| 久久精品91无色码中文字幕| 午夜久久久在线观看| 少妇猛男粗大的猛烈进出视频| 老司机在亚洲福利影院| 另类亚洲欧美激情| 男女边摸边吃奶| 国产精品一区二区免费欧美| 久久久久久免费高清国产稀缺| 国产精品九九99| 菩萨蛮人人尽说江南好唐韦庄| 久久精品亚洲熟妇少妇任你| 国产日韩欧美亚洲二区| 久久中文字幕一级| 久久久久视频综合| 国产精品免费视频内射| 少妇被粗大的猛进出69影院| 一级片免费观看大全| 国产成人一区二区三区免费视频网站| 黄色怎么调成土黄色| 男人舔女人的私密视频| 欧美日韩中文字幕国产精品一区二区三区 | 中文字幕最新亚洲高清| 一区二区三区激情视频| av视频免费观看在线观看| 纵有疾风起免费观看全集完整版| 国产精品久久电影中文字幕 | 伦理电影免费视频| 变态另类成人亚洲欧美熟女 | av欧美777| 电影成人av| 成人18禁在线播放| 777久久人妻少妇嫩草av网站| 精品乱码久久久久久99久播| 在线观看免费日韩欧美大片| 亚洲情色 制服丝袜| 在线亚洲精品国产二区图片欧美| 午夜福利在线免费观看网站| 国产区一区二久久| 欧美精品啪啪一区二区三区| h视频一区二区三区| 欧美亚洲 丝袜 人妻 在线| 人人澡人人妻人| 妹子高潮喷水视频| 两个人免费观看高清视频| 超色免费av| 国产成人免费无遮挡视频| 丝瓜视频免费看黄片| 久久中文字幕人妻熟女| 久久久久国产一级毛片高清牌| 欧美成人午夜精品| 欧美日本中文国产一区发布| 老司机午夜十八禁免费视频| 女性被躁到高潮视频| a级片在线免费高清观看视频| 亚洲色图 男人天堂 中文字幕| 亚洲av成人一区二区三| 免费黄频网站在线观看国产| 国产免费福利视频在线观看| 成年女人毛片免费观看观看9 | 精品国产一区二区久久| 99国产精品一区二区三区| 国产在线观看jvid| 电影成人av| 99精品在免费线老司机午夜| 亚洲成人免费av在线播放| 亚洲国产欧美一区二区综合| 人人澡人人妻人| 一个人免费看片子| 久久精品亚洲av国产电影网| 久久国产亚洲av麻豆专区| 乱人伦中国视频| 又大又爽又粗| 热re99久久国产66热| 欧美在线黄色| a在线观看视频网站| 男男h啪啪无遮挡| 精品视频人人做人人爽| 日本av免费视频播放| 欧美人与性动交α欧美软件| 久久中文看片网| av福利片在线| 色老头精品视频在线观看| 考比视频在线观看| 国产欧美日韩综合在线一区二区| 亚洲色图av天堂| 日本精品一区二区三区蜜桃| aaaaa片日本免费| 日韩一区二区三区影片| 国产男女超爽视频在线观看| 国产精品一区二区免费欧美| 精品人妻1区二区| 国产黄频视频在线观看| 国产av国产精品国产| 午夜福利乱码中文字幕| 美女主播在线视频| av电影中文网址| 久久久久精品人妻al黑| 桃花免费在线播放| 国产一区二区三区视频了| 一夜夜www| 大片免费播放器 马上看| 99精品欧美一区二区三区四区| 久久久久国产一级毛片高清牌| 日韩制服丝袜自拍偷拍| 青草久久国产| 色婷婷久久久亚洲欧美| 91精品三级在线观看| 桃花免费在线播放| 亚洲成人免费电影在线观看| 国产伦人伦偷精品视频| 亚洲精品久久午夜乱码| 他把我摸到了高潮在线观看 | 欧美 日韩 精品 国产| 视频在线观看一区二区三区| 欧美日韩国产mv在线观看视频| a级片在线免费高清观看视频| 老司机午夜福利在线观看视频 | 91精品三级在线观看| 午夜久久久在线观看| 国产成人系列免费观看| 欧美人与性动交α欧美精品济南到| av不卡在线播放| 视频区欧美日本亚洲| 色尼玛亚洲综合影院| 人人妻,人人澡人人爽秒播| 电影成人av| 亚洲精品国产色婷婷电影| 精品亚洲成a人片在线观看| 欧美精品一区二区免费开放| 1024香蕉在线观看| 男女无遮挡免费网站观看| 69精品国产乱码久久久| 50天的宝宝边吃奶边哭怎么回事| 男女午夜视频在线观看| 亚洲熟女毛片儿| 日本a在线网址| 这个男人来自地球电影免费观看| 丰满人妻熟妇乱又伦精品不卡| 建设人人有责人人尽责人人享有的| 最近最新中文字幕大全免费视频| 少妇猛男粗大的猛烈进出视频| 蜜桃在线观看..| 久9热在线精品视频| 日韩中文字幕欧美一区二区| 亚洲黑人精品在线| 久久午夜综合久久蜜桃| 国产日韩欧美在线精品| 国产黄色免费在线视频| 美女国产高潮福利片在线看| 一进一出好大好爽视频| 一级,二级,三级黄色视频| 一本久久精品| h视频一区二区三区| 久久国产精品大桥未久av| 桃花免费在线播放| 午夜精品国产一区二区电影| 国产亚洲精品久久久久5区| 91麻豆av在线| 国产男靠女视频免费网站| 中亚洲国语对白在线视频| 夜夜骑夜夜射夜夜干| kizo精华| 51午夜福利影视在线观看| 国产成人精品久久二区二区免费| 国产xxxxx性猛交| 色婷婷av一区二区三区视频| 手机成人av网站| 中文字幕人妻丝袜制服| 精品少妇久久久久久888优播| 嫩草影视91久久| 精品少妇黑人巨大在线播放| 精品一区二区三区四区五区乱码| 女人被躁到高潮嗷嗷叫费观| 日韩大码丰满熟妇| 黄色a级毛片大全视频| 亚洲,欧美精品.| 丰满人妻熟妇乱又伦精品不卡| 自线自在国产av| 国产熟女午夜一区二区三区| 热re99久久国产66热| 日韩熟女老妇一区二区性免费视频| 人人澡人人妻人| 丝袜在线中文字幕| 香蕉丝袜av| www.精华液| 热re99久久精品国产66热6| 久久精品成人免费网站| 国产高清国产精品国产三级| 久久亚洲精品不卡| 欧美日韩成人在线一区二区| 久久久久网色| 一二三四社区在线视频社区8| 久久午夜综合久久蜜桃| 亚洲av成人不卡在线观看播放网| 国产成人精品无人区| 色老头精品视频在线观看| 高清视频免费观看一区二区| av天堂久久9| av天堂在线播放| 无人区码免费观看不卡 | 一级黄色大片毛片| 在线天堂中文资源库| 高清av免费在线| 一边摸一边抽搐一进一出视频| 久久久久精品人妻al黑| 首页视频小说图片口味搜索| 妹子高潮喷水视频| 亚洲人成77777在线视频| 9191精品国产免费久久| 国产三级黄色录像| 久久久国产成人免费| 国产欧美日韩一区二区三区在线| 欧美成狂野欧美在线观看| 亚洲av片天天在线观看| 国产av精品麻豆| 一区二区三区乱码不卡18| 日韩视频一区二区在线观看| 免费久久久久久久精品成人欧美视频| av在线播放免费不卡| av网站在线播放免费| 少妇精品久久久久久久| 久久香蕉激情| 国产成+人综合+亚洲专区| 亚洲成人国产一区在线观看| 亚洲国产av新网站| 99国产极品粉嫩在线观看| 性少妇av在线| 在线观看免费视频日本深夜| 免费看十八禁软件| 麻豆成人av在线观看| 黄色视频,在线免费观看| 天天躁日日躁夜夜躁夜夜| 久久毛片免费看一区二区三区| 成年人黄色毛片网站| 国产亚洲精品一区二区www | www.自偷自拍.com| 一二三四社区在线视频社区8| 久久国产精品人妻蜜桃| 纯流量卡能插随身wifi吗| 久久人妻熟女aⅴ| 在线观看舔阴道视频| 亚洲久久久国产精品| 天堂8中文在线网| 99久久人妻综合| 麻豆av在线久日| 亚洲午夜精品一区,二区,三区| 午夜福利在线免费观看网站| 色综合欧美亚洲国产小说| 一二三四在线观看免费中文在| 岛国毛片在线播放| 国产99久久九九免费精品| 亚洲第一av免费看| 精品久久蜜臀av无| 亚洲精品国产区一区二| 国产成人精品久久二区二区91| 成年版毛片免费区| 国产精品二区激情视频| 欧美一级毛片孕妇| 天天影视国产精品| 国产伦人伦偷精品视频| 99九九在线精品视频| 欧美变态另类bdsm刘玥| 丰满人妻熟妇乱又伦精品不卡| 一边摸一边做爽爽视频免费| h视频一区二区三区| 国产精品成人在线| 国产片内射在线| 国产在线观看jvid| 大片电影免费在线观看免费| 99久久人妻综合| e午夜精品久久久久久久| 男男h啪啪无遮挡| 欧美黄色淫秽网站| 一个人免费看片子| 欧美日韩视频精品一区| 久久久久久人人人人人| 久9热在线精品视频| 久久精品国产综合久久久| 99国产综合亚洲精品| 熟女少妇亚洲综合色aaa.| 国产不卡av网站在线观看| 国产97色在线日韩免费| 黄色 视频免费看| 久久久国产成人免费| 在线观看舔阴道视频| 精品人妻1区二区| 麻豆国产av国片精品| 欧美精品亚洲一区二区| 51午夜福利影视在线观看| 男男h啪啪无遮挡| 欧美日韩亚洲综合一区二区三区_| 国产在线精品亚洲第一网站| 亚洲熟妇熟女久久| 亚洲欧洲精品一区二区精品久久久| 两个人免费观看高清视频| 国产精品秋霞免费鲁丝片| 黄色a级毛片大全视频| 亚洲精品粉嫩美女一区| 人人妻人人澡人人看| 麻豆乱淫一区二区| 最近最新中文字幕大全免费视频| 亚洲 欧美一区二区三区| 国产一区二区三区在线臀色熟女 |