• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Path Planning and Tracking Control for Parking via Soft Actor-Critic Under Non-Ideal Scenarios

    2024-01-27 06:48:20XiaolinTangYuyouYangTengLiuXiankeLinKaiYangandShenLi
    IEEE/CAA Journal of Automatica Sinica 2024年1期

    Xiaolin Tang ,,, Yuyou Yang , Teng Liu ,,,Xianke Lin , Kai Yang , and Shen Li ,,

    Abstract—Parking in a small parking lot within limited space poses a difficult task.It often leads to deviations between the final parking posture and the target posture.These deviations can lead to partial occupancy of adjacent parking lots, which poses a safety threat to vehicles parked in these parking lots.However, previous studies have not addressed this issue.In this paper, we aim to evaluate the impact of parking deviation of existing vehicles next to the target parking lot (PDEVNTPL) on the automatic ego vehicle (AEV) parking, in terms of safety, comfort, accuracy, and efficiency of parking.A segmented parking training framework(SPTF) based on soft actor-critic (SAC) is proposed to improve parking performance.In the proposed method, the SAC algorithm incorporates strategy entropy into the objective function, to enable the AEV to learn parking strategies based on a more comprehensive understanding of the environment.Additionally, the SPTF simplifies complex parking tasks to maintain the high performance of deep reinforcement learning (DRL).The experimental results reveal that the PDEVNTPL has a detrimental influence on the AEV parking in terms of safety, accuracy, and comfort, leading to reductions of more than 27%, 54%, and 26%respectively.However, the SAC-based SPTF effectively mitigates this impact, resulting in a considerable increase in the parking success rate from 71% to 93%.Furthermore, the heading angle deviation is significantly reduced from 2.25 degrees to 0.43 degrees.

    I.INTRODUCTION

    MANY collisions occur every year to the parking process due to the small parking space and complex driving skills [1].Autonomous vehicles are expected to safely and efficiently execute all driving skills [2], [3].However, In the field of intelligent vehicles, the technologies that can achieve batch application are mostly focused on advanced assisted driving and energy management [4], [5].Regarding the application of autonomous driving technology, its safety and reliability still need to be verified and tested [6], [7].automatic parking is one of the autonomous driving technologies, its safety, comfort, parking efficiency, and accuracy still need to be further studied.

    Currently, two types of automatic parking systems (APS)were proposed [8], [9]: one is general APS.After the system obtains the environmental information, it first plans a parking path.The vehicle is then controlled to follow the planned path to achieve parking.The other is the end-to-end APS, which combines the path planning module and path tracking control module.It directly outputs control variables (acceleration and steering angle) to control the vehicle according to environmental information.

    Currently, the primary path planning methods employed in general APS include quintic polynomial curve, cubic spline curve, cyclotron curve, and hybrid A-star algorithm, for tracking control, the commonly used algorithms are proportionintegral-derivative (PID), linear quadratic regulator (LQR),and model predictive control (MPC) [10]–[14].Some of them have been applied to advanced assisted driving systems.However, due to its ability to directly optimize the performance of the entire system, the end-to-end APS exhibits superior overall performance [15].The current methods of implementing end-to-end APS are deep learning and deep reinforcement learning (DRL) [16], [17].DRL is very useful in autonomous driving [18].It has great advantages not only in the end-to-end parking system but also in the parking lot allocation system[19]–[21].Some researchers have done many studies on it.For example, Zhanget al.[15] presented a DRL-based APS.The authors compared the parking performance between DRL-based APS and the general APS.The results proved that the DRL-based APS achieved park easily with fewer deviations.Songet al.[22] built a parking model based on actual parking data, which avoided the disadvantage that model-free RL requires a large amount of interactive data.The training efficiency was improved.However, The performance of the RL model is limited by this data.Bernhardet al.[23] used experience-based heuristic RL for automatic parking.Human main is used to weigh the data explored.Zhanget al.[24]treated parking as a multi-objective optimization process that considers safety, comfort, parking efficiency, and final parking performance.They obtained the parking control strategy considering all aspects through training the RL-based parking model.However, many difficulties still are not solved well in the practical applications of DRL-based APS.The gaps and weaknesses of previous research are that: Generalization is not enough good.The performance will decrease rapidly as the difficulty of the training task increases.The time of the training system model is too long.

    This paper proposes a segmented APS based on sofe actorcritic (SAC).It can improve the safety, comfort, and accuracy of automated parking and reduce training time by exploring more realistic environments and decomposing complex driving tasks into simple tasks.As shown in Fig.1, firstly, we quantify the parking deviation of existing vehicles next to the target parking lot (PDEVNTPL) on the automatic parking in terms of safety, comfort, efficiency, and accuracy in an experiment.Secondly, the best starting state for parking is determined to adapt to the target state that deviates from the ideal state due to PDEVNTPL, according to the target state, starting state, kinematic constraint, and PDEVNTPL.Thirdly, the best starting parking state is set as the segmentation point.By doing so, the complex parking task is effectively divided into a simpler parking task and a posture adjustment task.This decomposition reduces the overall difficulty of the parking process, addressing the issue of diminishing DRL performance as training tasks become more challenging.Following this, two driving tasks are trained using the SAC algorithm to acquire control strategies.SAC incorporates strategy entropy into the objective function, encouraging the vehicle to explore the environment and facilitating a more comprehensive learning process.Finally, the SAC-based segmented parking strategies are compared and analyzed against other algorithms.Safety, comfort, efficiency, and accuracy are evaluated using metrics such as success rate, trajectory smoothness, parking time, and posture deviation.

    The main contributions of this work are as follows: 1) The SAC algorithm is utilized to balance and improve the safety,comfort, efficiency, and accuracy of automatic parking.2)The impact of parking deviation of existing vehicles next to the target parking lot on the automatic ego vehicle parking is considered in the DRL training process, and it is quantified in terms of safety, comfort, and accuracy.3) A segmented parking training framework is established to simplify the parking task and collect the best parking starting state, to improve the generalization of APS.

    To better explain the contributions of this paper, the rest of the work is arranged as follows.Section II describes vehicle kinematics and parking scenarios.Section III introduces the SAC algorithm and segmented parking training framework,and implements the automatic parking system.Section IV analyzes the results.Section V presents a conclusion.

    II.VEHICLE KINEMATICS AND PARKING SCENARIO

    In this section, two different parking scenarios are introduced.The vehicle kinematics of automatic ego vehicle(AEV) is also described.The parking environment includes the AEV and surrounding vehicles.The surrounding vehicles were already parked.

    A. Vehicle Kinematics

    The rectangle is represented the outline of the vehicle.Its width and length are maximum width and length of the vehicle, which adds the restricted zone area to increase parking safety.As shown in Fig.2.x,yare the lateral position and longitudinal position of the vehicle, respectively.vis the vehicle velocity.lf,lris front wheelbase and rear wheelbase separately.ψis the heading.βis a slip angle at the center of gravity.δfis steering angle.Ris the turning radius.Since the vehicle travels at low speed during parking, the influence on driving by the physical property of tire and suspension is ignored.We can model the AEV by the bicycle model, which is based on the nonlinear continuous horizon equations [25], [26], as follows:

    The linear kinematic model is calculated by Taylor expansion

    For the convenience of the following research, the calculation formula of the minimum turning radius is given:

    The default parameters of the AEV and surrounding vehicles are the same.In this paper, the length and width of the vehicles are set to 5.0 meters and 2.0 meters respectively.The wheelbase is set to 3.6 meters (The front wheelbase equals the rear wheelbase).

    B. Parking Scenario

    Two types of parking scenarios are investigated in this study: the ideal parking scenario and the actual parking scenario.The actual parking scenario considers the parking deviations of existing vehicles next to the target parking lot (PDEVNTPL).As shown in Fig.3.The blue vehicle is the AEV.The gray vehicles are the surrounding vehicles.The green icon and red icon are the ideal target parking state and original target parking state.

    Fig.3.The ideal parking scenario and actual parking scenario.

    In actual parking scenario, the collision may happen if the AEV is parked according to the original target parking state.To coordinate the distances between AEV and adjacent vehicles.It is a good choice to adjust the target state by the mean of the PDEVNTPL.The basic change rules are as follows:

    where (?xT, ?yT, ?ψT) is the variation of the target parking

    state.?xi, ?yi, ?ψi, fori= 1, 2, are the lateral deviations, longitudinal deviations, and heading angle deviations of the adjacent vehicles respectively.For special cases where PDEVNTPL is a large value, which leads to the entrance of the parking lot being occupied, one extra rule is added: Ifmeters, reset the parking.

    According to China’s regulations on parking lot size, the length and width of mini parking lots are 5.5 meters and 2.5 meters.The width of the center lane of the parking lot should be 3.5 meters at least.Therefore, without losing generality, all parking environment sizes in this paper are set based on these regulations.To better reflect the influence of the PDEVNTPL,the width of the parking lot is changed to 3 meters.Take the target position as the origin.Right is the positive direction of theXaxis.Up is the positive direction of theYaxis.The lateral deviation (The abscissa difference between the centers of the vehicle and the parking lot), longitudinal deviations (The ordinate difference between the centers of the vehicle and parking lot) and heading deviations (The angle between the heading of a vehicle and the length direction of a parking lot)of the vehicles adjacent to the target parking lot take respectively a random value within [0, 0.5]/[-0.5, 0] (Vehicle on the left/Vehicle on the right) meters, [0, 0.2] meters and [-10, 10]degrees.(x,y,ψ) is used to represent the target state.The ideal target parking state is set as (0, 0, 90); The abscissa of the starting point should be greater thanRmin=6.4 (Equation (7)).For generality, the start parking state of the ideal parking scenario and general parking of the actual parking scenario is set as (8, 6, 0).The initial parking configuration of segmented parking in actual parking scenarios is introduced in Section III.

    III.IMPLEMENTATION OF AUTOMATIC PARKING SYSTEM

    In this section, the APS is established.Firstly, the SAC algorithm is introduced.Then, the segmented parking training framework is introduced.Finally, the implementation of the automatic parking system is introduced.

    A. SAC Algorithm

    In recent years, learning-based algorithms are a hot type of method in the field of intelligence [27]–[29].DRL is one type of learning-based algorithm.It has been applied in autonomous driving, mainly focusing on path planning, control strategy, and decision-making [30], [31].Deep deterministic strategy gradient (DDPG) algorithm is particularly outstanding due to its excellent performance [32].It is one of a few DRL algorithms that can solve the continuous action space problem.It was proposed in 2014 [33].However, Its training time is long.Its model generalization is not good.These problems need to be addressed urgently [34], [35].The SAC algorithm can overcome these shortcomings.It is a new DRL algorithm presented in 2018 [36].Currently, it is been applied in autonomous driving for decision-making [37] and path planning [38].In this study, we use it to learn parking control strategies.

    SAC is improved based on actor-critic.It borrows the experience replay mechanism of the deep Q-network to make the training data independent of each other.Also, it introduces strategy entropy to increase the randomness of action, which encourages vehicles to explore the environment.Strategy entropy is directly added into the Q function and value function to Maximize itself [39].

    where γ ∈(0,1) is the discounted factor to balance the shortterm and long-term reward.ρis a temperature parameter to tradeoff the entropy and reward.H(·|st) is entropy, expressed as

    As shown in Fig.4.SAC consists of five neural networks(NNs): one policy network is used to produce actions.one value network and one value-target network are used to evaluate the current states.Two Q-function networks are used to evaluate the effect of the action chosen based on the current state.

    Fig.4.The network structure and the update process of network parameters of SAC algorithm.

    The pseudo-code of SAC algorithm parameters training is displayed in Algorithm 1.The update of NNs parameters requires a certain amount of data.Before the formal training of the algorithm parameters, the vehicle randomly explores the environment to collect empirical data.After the data of replay bufferDis sufficient, the NNs parameters can be updated based on replay bufferD.

    Referring to reference [36], the updated principle of each network parameters of SAC algorithm is given below.The parameters of the value network are optimized by minimizing the value of the loss function.It is expressed as

    The gradient of the loss function is expressed as

    The value-target network has the same structure as the value network.Its parameters are optimized by copying from the value network periodically.

    For an algorithm with only one Q network (Such as deep Qnetwork, actor-critic), the evaluation of events is always overevaluated, which results in the optimization results being easily convergent to the local optimal.Van Hasseltet al.found that building two Q networks and selecting the smaller output of Q network as the target Q function during training, which can effectively alleviate the problem of algorithm overestimation [40].For the two Q-function networks, they have the same structure.Their parameters are optimized by minimizing soft Bellman residual function.It is expressed as

    whereQˉ(st,at) is the soft Bellman value of the state,expressed as

    where theVψˉ(st+1)is the output of the value target network.

    Algorithm 1 Soft Actor-Critic Ψ, ˉΨ, θ, ? 1: Initial parameters:2: for each iteration do 3: for each environment step do at ←π?(st,at)4: Select action:at st+1 rt 5: Execute ,observe next state and reward :st+1 ←p(st| st,at),rt=r(at,st)6: Store observes in replay memory D:D ←D∪(st,at,rt)7: end 8: for each gradient step do 9: Updata value network parameters:Ψ ←Ψ-λV ??ΨJv(Ψ)10:11: Updata the Q-function network parameters:θi ←θi-λQ ??θi JQ(θi) i ∈1,2, for 12: Updata policy network parameters:? ←?-λπ ???Jπ(?)13: Updata target network weights:ˉΨ ←τΨ+(1-τ)ˉυ 14: end 15: end

    The parameters of Q-function networks are optimized by following the gradient of soft Bellman residual function.It is expressed as:

    The parameters of the policy network are optimized by the loss function, expressed as

    The gradient of this loss function for optimizing parameters is expressed as

    where πtis the probability function of the distribution of actions.For continuous action space DRL, the probability function of the distribution of actions is used as the basis of action choice.Only the action with the highest probability is selected [36].It is expressed as

    whereμandσare the mean and standard deviation of the Gaussian distribution respectively.εis noise.Nmeans the standard normal distribution.The purpose of adding noise is to diffuse the range of action choices and increase vehicle exploration.

    B. Segmented Parking Training Framework

    The challenge in automatic parking lies in the limited freedom to adjust the vehicle’s posture within confined spaces,leading to difficulties in ensuring both safety and accuracy during the parking process.Additionally, the performance of DRL diminishes as the task complexity increases.Zhuanget al.[41] have proved that training efficiency can be improved by decomposing the parking process into three parking subprocesses.To address these issues, a segmented parking training framework (SPTF) is employed, as illustrated in Fig.5,providing a solution to these challenges.The complex parking task can be broken down into two simpler tasks: a posture adjustment task and a parking task.The best starting state (c2)for the simple parking task, evaluated by the Euclidean Distance between this state and the target state, can be calculated using the starting state (e2), target state, kinematic constraint,and PDEVNTPL.Two tasks are independently trained.By reducing the difference between the target state and the starting state, the SPTF enhances the available space for relative attitude adjustment.Therefore, The SPTF enables the AEV to effectively adjust its posture in crowded and limited spaces.

    Among the components of the Euclidean distance, the heading angle difference accounts for the largest proportion.Consequently, the primary goal when calculating the best starting state of parking (state c2) is to minimize the heading difference.As depicted in Fig.6(b), the maximum heading angle of the starting state of parking can be calculated using the following approach:

    wheredlaneis the width of center lane.LAEVandWAEVare the width and length of AEV.

    Substituting (24) into (23), φc2is obtained as follows:

    Fig.5.Segmentation and training framework of segmented parking based on DRL.

    Fig.6.Automatic parking trajectory planning based on the curve..

    The position ofd2is obtained as follow:

    Choosing the maximum allowed value for the heading angle φc2.The best starting state lateral position and longitudinal position are obtained as follows:

    whereLlotis the length of parking lot.

    xc2andyc2need to be adjusted according to the target state.Therefore, lateral positions and longitudinal positions of a2 and b2 are calculated based on variation of target state(?xT,?yT,?φT).

    C. Specific Implements

    1)The Implementation of APS Based on DRL: The implementation of using DRL to solve practical problems includes environment building and algorithm framework building.The environment is built in Section II.The algorithm framework will be built in this subsection.It generally includes five parts:confirming state variables and control variables, establishing reward function, building NNs construction, setting training mode and confirming training parameters.

    The state variable is the observation of AEV in the environment.It is generally expressed by position, velocity and heading angle.For more intuitive observation, this study projected each variable distribution onto two coordinate axes.Its expression is as

    where (xe,ye),ve, φeare the location coordinates, speed, and heading angle of the vehicle.(xT,yT),vT, φTare the location coordinates, speed, and heading angle of the target state.

    Control variables (action) is the command from the controller to AEV.This study confirms the control variables are accelerationaand front wheel steering angleδ.They are defined as follows:

    Parking safety (Parking without collisions), parking comfort (Control variables vary smoothly), parking efficiency(Parking trajectory is smooth), and deviations between final parking posture and target state are considered in this study.Therefore, the reward function consists of six parts corresponding to these four factors.It is expressed as

    whereRphconsiders the final parking posture.Rcfconsiders the parking comfort.Rvconsiders the parking efficiency.Rcis the collision penalty.Rsis the reward of success.The goal ofRadis used to reduce the lateral deviation of the parking.

    Rnis used to strengthen posture adjustment.Since the target state can change in actual parking scenario.TheRnin (43) is not feasible, changed as

    The NNs framework of SAC algorithm is shown in Fig.4.For two Q-function networks, their architectures are the same.Each network is composed of five layers: Layer 1 has 14 units to input the state and action information.Layers 2–4 is the hidden layer with 256 units.Layer 5 outputs the value of Q.

    For the policy network, value network and value-target network, their architectures are the same.Each network of they has five layers: Layer 1 has 12 units to input the state information.Layers 2-4 is the hidden layer with 256 units.Layer 5 outputs the mean and standard deviation of the action distribution and value of state.

    After much trial and error, the parameters of DRL are determined.The weight coefficients of the reward function and the hyper-parameters are displayed in Tables I and II.The simulation frequency is 15 Hz.The time of sampling action is 0.2 s.The duration of one episode is 50 steps or successful parking or collision.

    TABLE I THE HYPER-PARAMETERS OF SAC AND DDPG ALGORITHMS

    2)The Implementation of Motion Planning: Motion planning includes trajectory planning and speed planning.Safety,comfort, and accuracy are taken into account in trajectory planning.Speed planning takes comfort and driving efficiency into consideration.

    The trajectory planning method in this study is shown in Fig.6.The arc and straight line that meet the kinematic constraints are used to form a trajectory curve.It connects thestarting state to the target state.To solve the problem of curvature discontinuity, the cubic spline curve is used for interpolation fitting of the trajectory.The trajectory for the segmented parking (as shown in Fig.5(b)) is calculated using (23) to(28).The equations of the trajectory of general parking (Fig.5(a)) are the same as the trajectory of segmented parking.The following changes are made:dlane, φc2is set asysand 0.5π.ysis the longitudinal position of starting position.

    TABLE II THE WEIGHT COEFFICIENT OF REWARD FUNCTION

    The principle of cubic spline curve interpolation fitting is to solve the cubic polynomial function between adjacent nodes.It is represented as follows:

    By substituting the boundary values, (45) is derived to obtain the solution

    wherehi=[xi+1-xi].mi,i∈[0,1,...,N] is the solution to the following equation:

    Comfort is considered in the objective function of the algorithm.Speed planning is as follows:

    wherek,k∈{-1,1}, is directional coefficient.dis the distance between AEV and starting position.d0is the distance between starting position and target position.vmaxis the maximum speed allowed.

    Parameters of motion planning are set as follows: Trajectory sampling interval is 0.1 meters; Maximum allowable speedvmax= 3 m/s.The boundary conditions of cubic spline curve is Natural Spline.

    3)The Implementation of Tracking Control Based on MPC:The MPC algorithm is known for its simulation prediction capabilities.It aims to find a set of optimal solutions within the prediction time horizon through iterative optimization.The linear system model exhibits a favorable prediction performance.Therefore, the implementation steps for MPC path tracking involve the linearization of the tracking control model and the construction of the objective function.

    The linear equation of the approximate system can be obtained by the linear kinematic model (Equation (6))

    where

    The purpose of the objective function is to reduce the tracking error.Comfort is also considered in this study.The objective function is set as follows:

    where theQMPC,RMPC,Rdare weight matrixes.Nis the number of simulation steps of prediction (Predictive time domain).

    Aftermuchtrialand error,theMPC parameters aresetas follows:N=30,QMPC=[2211]T,RMPC=[0.10.1]T,Rd=[0.01 0.01]T.The maximum allowable loss value is 0.01.Maximum iteration step is 100.

    4)The Implementation of Tracking Control Based on LQR and PID: Previously, researchers have explored the idea of breaking down a multi-input tracking control system into multiple single-input tracking control systems by decoupling the control variables [42].However, it should be noted that a complete tracking control system is still interconnected through mutual state variables.One common approach involves designing the path tracking control system and the speed tracking control system separately, with the overall tracking control system being coupled through the velocity variable.In this study, the path tracking control system is designed using the LQR method, while the speed tracking control system is designed using the PID controller.

    LQR is also used to solve linear system problems.A linear model for tracking control is given as follow:

    where

    de(t) and φe(t) are position error and heading angle error respectively.

    The cost function is defined as follows:

    whereQLQR,RLQRare weight matrixes.Mis predictive time domain.

    The solution of LQR is as follows:

    Substituting (44) into (43), takeM=∞, the equation forKis obtained as follows:

    wherePis the solution to the following Riccati equation:

    PID is used to track control speed.Its formula is expressed as follows:

    wheree(t) is function of error.

    After many trials and errors, the parameters of PID and LQR are determined as follows:QLQR=eye[100 0.10 0.1],RLQR=[0.1].The maximum number of iterations is 150.The maximum allowable loss value is 0.01.The values ofkp,ki,andkdare set to 2, 0.001, and 0.1 respectively.

    5)The Implementation of APS Based on DDPG And OBCA:The basic trajectory is planned based on DDPG, and then the trajectory is optimized by optimization-based collision avoidance (OBCA).The safe and comfortable parking trajectory is obtained [43].The implementation process of DDPG is the same as that of SAC (30-44), which will not be repeated here.

    The implementation of the OBCA-based optimization of the path planned by DDPG include the construction of objective functions and the addition of safe constraints.The basic of trajectory optimization is the basic trajectory.Therefore, the objective function and path tracking are similar.It is expressed as follows:

    whereQ,Rare weight matrixes.Their value selection is the that of the MPC.

    xrefis the points sampled from the basic trajectory.xanduis the state variables and control variables.Mis the number of points sampled.

    The constraint types are kinematic constraint, obstacle avoidance constraint, control quantity range constraint and trajectory starting and ending constraint.It is expressed as follows:

    wherexstartandxtargetare the start point and target point of parking trajectory.f(xk,uk) if the vehicle kinematics model(6).(o1, o2, o3, o4) and (v1, v2, v3, v4) are four corners of a rectangle representing vehicles and obstacles.S?is triangular area.Sis rectangular area.

    IV.EXPERIMENTS AND DISCUSSION

    This section evaluates the performance of SAC-based segmented parking strategy.The simulation environment is constructed based on Python 3.7 [44].Firstly, the convergence of DRL training strategies is analyzed.Secondly, the performance of AEV parking is compared between DRL methods and traditional methods.Thirdly, the influence of PDEVNTPL on automatic parking is quantified in terms of safety,comfort, efficiency, and accuracy.Finally, The effect of that SPTF is used to reduce the influence of PDEVNTPL is analyzed.

    A. Policy Convergence

    In the subsection, the training process is analyzed.The SAC and DDPG algorithms are to train AEV to park Safely, comfortably, efficiently and accurately.The learning effect is reflected by the cumulative reward of an episode.AEV must learn to balance the four elements to maximize the accumulative reward.In the training process, the parking policy is improved by increasing of the accumulative reward.When the cumulative reward converges to a value with small fluctuations, the parking policy has reached the optimum.

    The automatic ego vehicle is trained in over 100 000-time steps (2000 episodes).The average reward of the training process is shown in Fig.7.AEV is trained by SAC and DDPG.Both cumulative reward curves tend to rise first and converge to a value finally.Obviously, the cumulative reward of SAC(600 episodes) becomes convergence earlier than DDPG(1800 episodes).The convergence values of the two algorithms are different little.It means SAC has a higher training efficiency while ensuring good learning effects.

    Fig.7.The average reward for SAC and DDPG training process.

    According to the analysis of the above experimental results,SAC algorithm can greatly improve the training efficiency,while ensuring the same convergence performance as DDPG.

    B. Comparison of Parking Performance of Different Methods

    In this subsection, the performances of DRL methods, traditional methods, and scheme combining DRL and traditional methods on parking are compared in the ideal parking scenario.DRL methods are SAC algorithm and DDPG algorithm[15].Traditional methods are that Spline curve and MPC are used to plan path and track control (S+MPC) [12], and Spline curve, LQR and PID are used to plan path and track control(S+LQR+PID) [11].scheme combining DRL and traditional methods is that DDPG is used to plan path and OBCA is used to correct the trajectory and track the control (DDPG+OBCA)[43].The performance of each method is analyzed in terms of safety, comfort, efficiency, and accuracy.

    The metrics used to evaluate automated parking systems are generally safety, comfort, efficiency, and accuracy [15],[22]–[24].Safety is concerned with determining if any collisions occur during the parking process.Comfort assesses whether the control variables, namely acceleration and steering angle, experience frequent and significant fluctuations.Efficiency measures the duration of the parking process.Accuracy evaluates the deviation between the final parking state and the desired target state.In the ideal parking scenario,the AEV would be able to park without any collisions consistently.Consequently, the analysis of APS will focus on the comfort, efficiency, and accuracy aspects, examining factors such as parking deviations, control variable smoothness, and parking time.

    The reward function assigns varying weights to different criteria, with accuracy being accorded only smaller significance than safety.This weighting suggests a prioritization of optimizing accuracy initially, under the condition of ensuring safety.The parking deviations of all methods are tabulated in Table III.All methods adhere to BS ISO16787 standards, barring the S+MPC level deviation.Notably, the DRL method demonstrates reduced deviations in comparison to the traditional methods.For instance, the S+MPC lateral deviation and the heading deviation of S+LQR+PID is more pronounced.Conversely, deviations in the case of DDPG and SAC remain relatively modest, both in positional deviation (below 0.1 meters) and heading angle deviation (below 1 degree).The deviation in DDPG+OBCA is comparable to, or even smaller than that of the DRL method.The performance of the DDPG technique is notably commendable.Post reoptimization, further enhancement in system performance is achieved.

    TABLE III THE PARKING DEVIATIONS IN IDEAL PARKING SCENARIO AND BS ISO16787 FOR COMPARISON

    Fig.8 shows the control variables of parking in the ideal parking scenario.To directly observe the comfort of parking,the smoothness of the control variable is calculated using the following equation [41]:

    whereaˉ is the average value of the set of control variables.aiis the element of the set of control variables.nis the length of the control variables set.

    Fig.8.The acceleration and steering angle for automatic parking in the ideal parking scenario.

    TABLE IV THE SMOOTHNESS OF ACCELERATION AND STEERING ANGLE IN IDEAL PARKING SCENARIO

    Table IV presents the smoothness characteristics of the control variables.Generally, both S+MPC and DDPG+OBCA exhibit superior smoothness (lower values) when contrasted with the two DRL methods.S+LQR+PID displays the least favorable smoothness (maximum values), particularly evident in the steering angle domain.This observation aligns logically with the fact that model-based methods (MPC and OBCA) can adeptly fine-tune parameters based on the model,excelling in individual aspects.However, a challenge arises in finding an encompassing parameter set ensuring holistic optimization.LQR enhances real-time performance by maximizing subsequent state returns, which limits its agility in coping with abrupt state transitions.Instances of sharp curvature fluctuations necessitate substantial control variable adjustments for minimal tracking error.Conversely, DRL optimizes policies by considering both immediate actions and their downstream consequences, preserving global performance optimality.Consequently, the relative reduction in comfort within DRL compared to model-based methods (MPC and OBCA) is attributed to DRL’s primary training focus on secure and precise parking, with comfort inadvertently sacrificed.This compromise is further compounded in the pursuit of parking efficiency optimization, causing a detriment to comfort within the DRL methods.As a result, DRL exhibits suboptimal performance in the comfort aspect.

    Parking efficiency and the smoothness of acceleration are often conflicting objectives.Maximizing efficiency requires aggressive acceleration and deceleration, leading to significant changes in acceleration during the parking process.As a result, the smoothness of acceleration is unavoidably affected.The parking times is presented in Table V.The DRL methods prioritize the reduction of parking time over maintaining smooth acceleration, thereby sacrificing trajectory smoothness.For instance, the parking duration for the DDPG method is 7 seconds, while after the incorporation of DDPG+OBCA,the parking duration extends to 13.6 seconds.OBCA substantially enhances the smoothness of the control variables, resulting in a considerable reduction in parking efficiency.

    Based on experimental analysis, it is evident that the DRL method outperforms traditional approaches in terms of auto-mated parking performance.When compared with the stateof-the-art methods, precision levels can achieve high standards, albeit at the cost of reduced comfort, but with a notably greater efficiency.While the DRL method might not offer optimal comfort, this trade-off is made to attain elevated levels of efficiency and accuracy.

    TABLE V THE PARKING TIME OF AUTOMATIC PARKING IN IDEAL PARKING SCENARIO

    C. Quantification Influence of PDEVNTPL on Automatic Parking

    PDEVNTPL is a common occurrence in real-life situations,and it significantly affects automatic parking.This is clearly illustrated in Fig.9, where the PDEVNTPL poses a threat to the successful execution of automatic parking.In this subsection, the impact of PDEVNTPL on automatic parking is quantified through tests conducted in actual parking scenarios.

    Fig.9.The influence of PDEVNTPL in automatic parking in actual parking scenario.(a) The rear of the vehicle takes up parking space and causes a collision; (b) The head of the vehicle blocks the entrance by occupying the parking space.

    During the training of DRL, the horizontal deviation of adjacent vehicles is randomly set between 0 and 0.5 meters,while the heading angle deviation is randomly set between -10 and 10 degrees.It is observed that the influence of heading angle deviation is more significant.During testing, the heading angle deviation also remains random within the range of -10 to 10 degrees, to ensure a continuous influence on automatic parking.Additionally, the horizontal deviation increases sequentially.To study the extent of influence from PDEVNTPL, the parking space is gradually reduced, and the parking difficulty is increased.The success rate, average smoothness of control variables, and average deviations of the final state are computed based on 1000 test results.These metrics serve to quantify the degree of influence PDEVNTPL on automatic parking.

    The success rates of the experiments are presented in Table VI.As the lateral deviation increases, the parking space gradually diminishes, leading to heightened parking difficulty.Consequently, the success rates of all methods exhibit a decline.At the maximum PDEVNTPL value (0.5), the success rates of the S+MPC and S+LQR+PID algorithms plummet by over 30%.Similarly, the success rates of SAC and DDPG decrease by 29%.Even for DDPG+OBCA, whichenforces safety constraints for assurance, the success rate experiences an almost 10% drop.Safety constraints ensure that the target vehicle avoids collisions with neighboring vehicles but do not guarantee safety throughout the entire trajectory from start to finish.Given the confined parking space,certain scenarios necessitate conditions that prevent collisions with adjacent vehicles, inadvertently leading to collisions with vehicles in upper parking spaces.

    TABLE VI THE SUCCESS RATE OF GENERAL PARKING WITH VARIATION OF DEVIATION

    Parking deviations are presented in Table VII.All deviations with PDEVNTP exceed these deviations without PDEVNTP (Table III).The influence of adjacent vehicle parking deviations on the autonomous parking of the AEV is primarily reflected in heading angle deviations and lateral deviations.The lateral deviations for S+MPC, S+LQR+PID, DDPG,DDPG+OBCA, and SAC increased by 230%, 153%, 300%,200%, and 359% respectively.Similarly, the heading angle deviations increased by 196%, 21%, 32%, 101%, and 142%respectively.These deviations are primarily impacted by PDEVNTPL.Notably, the influence on longitudinal deviation is minimal.This is attributed to the increased complexity in posture adjustments stemming from alterations in the target state, which has no bearing on velocity planning.

    TABLE VII THE PARKING DEVIATIONS OF GENERAL PARKING IN ACTUAL PARKING SCENARIO

    Regarding parking comfort, the smoothness of control variables is presented in Table VIII.In comparison to parking without PDEVNTPL, all methods exhibit poorer (higher values) steering angle smoothness (Table IV), with relatively minor variations in acceleration smoothness.Fig.10 provides an intuitive understanding of the results.To effectively accommodate deviations between the target state and desired state, the steering Angle requires a large adjustment.Inversely, acceleration necessitates only minor adjustments to accommodate distance variations between the starting and tar-get points.The steering angle smoothness for S+MPC,S+LQR+PID, DDPG, DDPG+OBCA, and SAC increased by 27%, 6%, 9%, 72%, and 58% respectively, indicating a decline in smoothness.Notably, both S+LQR+PID and DDPG exhibit minimal variations in steering angle smoothness due to their poor comfort levels, as evident in Table IV, with a decrease in the degree of variation as they approach worstcase scenarios.

    TABLE VIII THE ACCELERATION AND STEERING ANGLE SMOOTHNESS OF GENERAL PARKING IN ACTUAL PARKING SCENARIO

    Fig.10.The acceleration and steering angle for general parking in actual parking scenario.

    Based on the aforementioned analysis of experimental data,PDEVNTPL exerts an impact on automated parking across all methods.It primarily influences success rates, lateral deviation, heading angle deviation, and steering angle smoothness.Under the influence of PDEVNTPL, safety, accuracy, and comfort are reduced by approximately 27%, 54%, and 26%respectively.

    D. The Effect of Reducing The Impact of PDEVNTPL by SPTF

    In this subsection, the SPTF is employed to minimize the impact of PDEVNTPL in actual parking scenarios.The path planning of traditional methods also follows the principles of the SPTF.By conducting a comparative analysis of parking safety, accuracy, comfort, and efficiency between general parking and SPTF-based parking, the advantages of the SPTF are verified.To ensure an accurate evaluation of parking strategy performance, data from 1000 tests are collected, and the average values are used as the main criteria.

    In terms of parking safety, the success rates are presented in Table IX.The experimental results indicate that success rates decrease with increasing PDEVNTPL values.However, even when PDEVNTPL reaches its maximum (0.5), the success rate of the SAC method remains above 93%, surpassing the success rates of both DDPG and DDPG+OBCA.The successrates of traditional methods also exhibit improvement (Over 20% improvement), albeit not as prominently as SAC.The direct cause of enhanced parking safety is the SPTF, it ensures the AEV initiates parking in the most best state.As depicted in Fig.11, the initial orientation angle is usually 90 degrees different from the target orientation angle in general parking scenarios.Additionally, there is a larger lateral distance to cover.SPTF ensures that the initial state is not only closer to the target state but also possesses a smaller angular difference.This allows the AEV to achieve parking through simple adjustments.Moreover, the value of the best starting state for parking, as determined by SPTF, takes PDEVNTPL into account, enabling the AEV to adapt to non-ideal target states induced by PDEVNTPL.

    TABLE IX THE SUCCESS RATE OF SPTF-BASED PARKING AND GENERAL PARKING WITH VARIATION OF DEVIATION FOR COMPARISON

    Fig.11.General parking and SPTF-based parking processes.(a) General parking is unable to adjust the attitude due to the large distance between the starting state and the target state, resulting in a collision; (b) The distance between the starting state and the target state of SPTF-based parking is small,AEV parks easily.

    In terms of accuracy, the parking deviations are presented in Table X.It can be observed that the parking deviations are all smaller than those of general parking.Lateral deviation reduced to the level of parking without PDEVNTPL (Table IV).However, it is noteworthy that there still exists a relatively large deviation for these four methods.Nevertheless,the success rate of the SAC algorithm surpasses that of the other algorithms.This implies that even under larger PDEVNTPL values, SAC is capable of achieving parking safely.These cases present greater challenges to the parking task,resulting in larger deviations.SAC prioritizes safety over other performance aspects, sacrificing them to ensure a higher level of safety during parking.

    TABLE X THE PARKING DEVIATIONS OF SPTF-BASED PARKING AND GENERAL PARKING IN ACTUAL PARKING SCENARIO FOR COMPARISON

    In terms of parking comfort, the smoothness of control variables is illustrated in Table XI.It can be observed that, in comparison to general parking, the smoothness of all methods is enhanced (values are lower), particularly in terms of steering angle smoothness.By employing the SPTF, the parking task is divided into two phases: posture adjustment and simple parking.While the overall amount of posture adjustment remains constant, the total amount of posture adjustment for individual tasks decreases.This partitioning enables a smoother and more comfortable parking experience.The best starting state further reduces the overall amount of posture adjustment required, consequently minimizing abrupt changes in control variables.As depicted in Fig.12, it is evident that the smoothness of SAC is inferior (higher values) compared to S+MPC and DDPG+OBCA.This disparity can be attributed to SAC’s primary focus on optimizing safety during the parking process.Comfort is further sacrificed to enhance parking efficiency.

    TABLE XI THE ACCELERATION AND STEERING ANGLE SMOOTHNESS OF SPTF-BASED PARKING AND GENERAL PARKING IN ACTUAL PARKING SCENARIO FOR COMPARISON

    Fig.12.The acceleration and steering angle of SPTF-based parking in actual parking scenario.

    TABLE XII THE PARKING TIME OF SPTF-BASED PARKING AND GENERAL PARKING FOR COMPARISON

    The parking durations based on SPTF are presented in Table XII.Experimental results demonstrate that the parking times for most methods increase by more than 200% compared to general parking times.This is a drawback of the SPTF approach.However, sacrificing some efficiency in favor of enhanced safety and accuracy is reasonable.Among these methods, S+MPC, DDPG, and SAC exhibit shorter parking durations compared to S+LQR+PID.SAC achieves higher success rates.As the deviation of the target state increases, the parking duration also increases.Naturally, if parking speeds are the same, the parking duration for S+MPC would naturally be longer.Overall, SAC performs better in terms of parking duration.

    Through the analysis of experimental data, it can be concluded that the SPTF effectively mitigates the impact of PDEVNTPL on automated parking, leading to a notable enhancement in automated parking performance.The parking success rate exceeds 90%, and parking deviations are reduced to a level comparable to parking without PDEVNTPL.There is also a certain degree of improvement in comfort.Among the tested algorithms, SAC demonstrates the most pronounced performance improvement.

    V.CONCLUSION

    In order to assess the impact of parking deviation caused by other vehicles on automatic parking, we conducted a quantitative analysis.It was observed that the PDEVNTPL significantly affects various aspects of automatic parking performance, with safety, comfort, and accuracy experiencing a decline of more than a quarter.To mitigate this influence, we propose a SAC-based SPTF approach, which successfully reduces the impact of PDEVNTPL.Consequently, the safety,comfort, and accuracy of parking are substantially improved.However, it is important to note that the DRL-based APS used in this study is not suitable for unfamiliar scenarios due to the inherent characteristics of RL and the specificity of the reward function.Moreover, DRL methods are unable to incorporate state or control variable constraints to ensure vehicle safety in abnormal conditions.Furthermore, the study did not consider the influence of moving obstacles near the target parking lot or the impact of incomplete environmental information resulting from sensor failures on automatic parking.These factors will be addressed in future research, which will focus on autonomous parking scenarios where ego vehicles interact with nearby vehicles.Additionally, the study will explore the safety implications of varying loss ratios, types, and distances of environmental information.Finally, the parking control strategy will be transferred to real-world vehicles for experimental validation.

    国产亚洲欧美98| 国产精品日韩av在线免费观看| 欧美日韩精品网址| 亚洲午夜理论影院| 丁香六月欧美| 久久久久精品国产欧美久久久| 别揉我奶头~嗯~啊~动态视频| 成人无遮挡网站| 99久久无色码亚洲精品果冻| 成人特级av手机在线观看| 人人妻人人澡欧美一区二区| 日本熟妇午夜| 我要搜黄色片| 亚洲在线观看片| 草草在线视频免费看| 特级一级黄色大片| 国产成+人综合+亚洲专区| 波多野结衣巨乳人妻| 国产伦精品一区二区三区四那| 夜夜躁狠狠躁天天躁| 国产av一区在线观看免费| 欧美黄色淫秽网站| 国内精品久久久久久久电影| x7x7x7水蜜桃| 免费看日本二区| 色综合欧美亚洲国产小说| 性欧美人与动物交配| 99热6这里只有精品| 久9热在线精品视频| 99riav亚洲国产免费| 欧美zozozo另类| 午夜亚洲福利在线播放| 色综合亚洲欧美另类图片| 12—13女人毛片做爰片一| 黄色 视频免费看| 一进一出抽搐gif免费好疼| 国产高清视频在线播放一区| 午夜两性在线视频| 不卡av一区二区三区| 亚洲一区二区三区不卡视频| 一个人看的www免费观看视频| 1024香蕉在线观看| 国产精品久久久久久精品电影| 国产69精品久久久久777片 | 99re在线观看精品视频| 97超视频在线观看视频| 亚洲国产精品合色在线| 天堂网av新在线| 国产 一区 欧美 日韩| 一边摸一边抽搐一进一小说| 他把我摸到了高潮在线观看| 男人舔女人下体高潮全视频| 91av网一区二区| 男女午夜视频在线观看| 日韩欧美 国产精品| 精品福利观看| 国产精品一区二区三区四区免费观看 | 亚洲av成人精品一区久久| 国内精品久久久久久久电影| 国产真实乱freesex| 久久午夜亚洲精品久久| 亚洲成av人片免费观看| 十八禁人妻一区二区| 91久久精品国产一区二区成人 | 草草在线视频免费看| 亚洲成人中文字幕在线播放| 女人高潮潮喷娇喘18禁视频| 99久久久亚洲精品蜜臀av| 熟女电影av网| 国产免费男女视频| 高潮久久久久久久久久久不卡| 精品欧美国产一区二区三| 在线视频色国产色| 亚洲av片天天在线观看| 久久亚洲真实| 色综合亚洲欧美另类图片| av天堂在线播放| 在线观看66精品国产| 国产精品亚洲美女久久久| 亚洲在线观看片| www.精华液| 久99久视频精品免费| 两性夫妻黄色片| 99国产精品一区二区蜜桃av| 国模一区二区三区四区视频 | 97超视频在线观看视频| 欧美色视频一区免费| 精品久久久久久久久久免费视频| 欧美日韩亚洲国产一区二区在线观看| 久久精品影院6| 女人被狂操c到高潮| 国产一区二区在线av高清观看| 少妇熟女aⅴ在线视频| 久久国产精品人妻蜜桃| 99久久成人亚洲精品观看| 精品午夜福利视频在线观看一区| x7x7x7水蜜桃| 黑人操中国人逼视频| av天堂中文字幕网| 久久久久久久久久黄片| 久久久色成人| 欧美日本视频| 一a级毛片在线观看| 国产黄色小视频在线观看| 国产成人av教育| 天堂av国产一区二区熟女人妻| 国产一级毛片七仙女欲春2| 亚洲,欧美精品.| 日本一二三区视频观看| 丁香六月欧美| 久久久国产欧美日韩av| 免费av毛片视频| 我的老师免费观看完整版| 亚洲精品国产精品久久久不卡| 伊人久久大香线蕉亚洲五| 精品不卡国产一区二区三区| 国产精品久久久久久久电影 | 久久久久精品国产欧美久久久| 美女 人体艺术 gogo| 国产激情欧美一区二区| 日韩免费av在线播放| 亚洲精品国产精品久久久不卡| 日韩有码中文字幕| 国产精品99久久久久久久久| 热99在线观看视频| 欧美成狂野欧美在线观看| 精品一区二区三区av网在线观看| 免费在线观看影片大全网站| 叶爱在线成人免费视频播放| 99热这里只有精品一区 | 极品教师在线免费播放| 国产精品 欧美亚洲| 在线观看66精品国产| 国产精品免费一区二区三区在线| 亚洲人成电影免费在线| 免费在线观看视频国产中文字幕亚洲| 国产伦精品一区二区三区视频9 | 老司机午夜十八禁免费视频| 欧美日韩瑟瑟在线播放| 国产精品免费一区二区三区在线| 美女 人体艺术 gogo| 亚洲一区二区三区不卡视频| 国产伦精品一区二区三区视频9 | bbb黄色大片| 中文字幕高清在线视频| 啦啦啦免费观看视频1| 岛国在线免费视频观看| 亚洲男人的天堂狠狠| 亚洲欧美精品综合一区二区三区| 国产午夜精品久久久久久| 在线观看午夜福利视频| 黄色视频,在线免费观看| 国产伦精品一区二区三区四那| 看免费av毛片| 欧美另类亚洲清纯唯美| aaaaa片日本免费| 久久久久久久精品吃奶| 国产精品av久久久久免费| 国产一区二区在线av高清观看| 长腿黑丝高跟| 亚洲精品国产精品久久久不卡| 久久精品国产亚洲av香蕉五月| 国产精品久久久人人做人人爽| 亚洲精品美女久久久久99蜜臀| 后天国语完整版免费观看| 精品久久久久久久久久免费视频| 午夜影院日韩av| 欧美激情在线99| 国产aⅴ精品一区二区三区波| 夜夜看夜夜爽夜夜摸| 五月玫瑰六月丁香| 少妇的丰满在线观看| 成年版毛片免费区| 亚洲国产欧美人成| 国产激情久久老熟女| 亚洲人成网站在线播放欧美日韩| 亚洲,欧美精品.| 国产麻豆成人av免费视频| 亚洲国产欧美网| 在线视频色国产色| 亚洲欧美日韩高清专用| 成人永久免费在线观看视频| 一进一出好大好爽视频| x7x7x7水蜜桃| 黄色成人免费大全| 国产一区二区三区在线臀色熟女| 99久久成人亚洲精品观看| 每晚都被弄得嗷嗷叫到高潮| 最近最新中文字幕大全免费视频| 国内精品久久久久精免费| av中文乱码字幕在线| av视频在线观看入口| 国产aⅴ精品一区二区三区波| 精品电影一区二区在线| 久久精品91无色码中文字幕| 成人欧美大片| 午夜免费成人在线视频| 亚洲精华国产精华精| 夜夜躁狠狠躁天天躁| 神马国产精品三级电影在线观看| 嫁个100分男人电影在线观看| 淫妇啪啪啪对白视频| 12—13女人毛片做爰片一| 我要搜黄色片| 天堂av国产一区二区熟女人妻| 少妇丰满av| 欧美色视频一区免费| 又黄又粗又硬又大视频| 久久久水蜜桃国产精品网| 美女 人体艺术 gogo| 亚洲欧美精品综合一区二区三区| 精品国产乱码久久久久久男人| 精品久久久久久久久久久久久| 男人舔奶头视频| 国模一区二区三区四区视频 | 99精品欧美一区二区三区四区| 亚洲成a人片在线一区二区| 又紧又爽又黄一区二区| 精品国产乱码久久久久久男人| 黑人欧美特级aaaaaa片| 久久亚洲精品不卡| 国产亚洲欧美98| 亚洲成人免费电影在线观看| 欧美成人性av电影在线观看| 波多野结衣巨乳人妻| 禁无遮挡网站| 岛国在线观看网站| 俄罗斯特黄特色一大片| 日日摸夜夜添夜夜添小说| 日本 av在线| 一进一出抽搐动态| 夜夜看夜夜爽夜夜摸| 欧美日韩乱码在线| 亚洲精品乱码久久久v下载方式 | 在线观看美女被高潮喷水网站 | 青草久久国产| 两性午夜刺激爽爽歪歪视频在线观看| 露出奶头的视频| 丰满人妻熟妇乱又伦精品不卡| 此物有八面人人有两片| 男人和女人高潮做爰伦理| 久久精品亚洲精品国产色婷小说| 国产亚洲欧美在线一区二区| 国内久久婷婷六月综合欲色啪| 999精品在线视频| 日本a在线网址| 成人欧美大片| 五月玫瑰六月丁香| 伦理电影免费视频| www日本黄色视频网| 成人av在线播放网站| 天天一区二区日本电影三级| 18禁黄网站禁片免费观看直播| 国产精品av久久久久免费| 可以在线观看的亚洲视频| 免费在线观看日本一区| 久久精品91蜜桃| 午夜免费观看网址| 午夜精品在线福利| 午夜日韩欧美国产| 国产一区在线观看成人免费| 成人午夜高清在线视频| 久久久久国产一级毛片高清牌| 亚洲国产日韩欧美精品在线观看 | 又粗又爽又猛毛片免费看| 麻豆国产97在线/欧美| 1024手机看黄色片| 怎么达到女性高潮| 国产高清视频在线播放一区| 国产精品一区二区三区四区久久| 男人和女人高潮做爰伦理| 日韩大尺度精品在线看网址| 12—13女人毛片做爰片一| 欧美性猛交黑人性爽| 国内精品久久久久久久电影| 亚洲va日本ⅴa欧美va伊人久久| 久久中文看片网| 首页视频小说图片口味搜索| 国产精品久久久久久亚洲av鲁大| 一级作爱视频免费观看| 亚洲一区二区三区色噜噜| a级毛片在线看网站| 9191精品国产免费久久| 国产一区二区在线av高清观看| 国产av一区在线观看免费| 国产野战对白在线观看| 久久这里只有精品19| 国产一区二区三区视频了| 免费搜索国产男女视频| 99国产精品一区二区三区| 亚洲中文av在线| 亚洲人成网站高清观看| 在线永久观看黄色视频| 成人18禁在线播放| 亚洲片人在线观看| 成人鲁丝片一二三区免费| 久久久久久人人人人人| 毛片女人毛片| 99视频精品全部免费 在线 | 18禁观看日本| 成年人黄色毛片网站| 免费无遮挡裸体视频| 香蕉丝袜av| 一区二区三区国产精品乱码| 长腿黑丝高跟| 国产探花在线观看一区二区| 99精品欧美一区二区三区四区| svipshipincom国产片| 精品久久久久久,| 午夜激情福利司机影院| 国产一级毛片七仙女欲春2| 精品久久久久久久末码| 成人高潮视频无遮挡免费网站| 久久天堂一区二区三区四区| 无人区码免费观看不卡| 日本一本二区三区精品| 中文亚洲av片在线观看爽| 琪琪午夜伦伦电影理论片6080| 黄色片一级片一级黄色片| 人妻丰满熟妇av一区二区三区| 免费av不卡在线播放| av欧美777| 中国美女看黄片| 午夜精品一区二区三区免费看| 此物有八面人人有两片| 99国产精品一区二区蜜桃av| 91老司机精品| 村上凉子中文字幕在线| 成人无遮挡网站| 成人精品一区二区免费| 国产一区二区在线观看日韩 | 最近视频中文字幕2019在线8| 校园春色视频在线观看| 亚洲一区二区三区色噜噜| 国产精品女同一区二区软件 | 在线观看66精品国产| 一个人免费在线观看的高清视频| svipshipincom国产片| 欧美在线黄色| 色综合欧美亚洲国产小说| 成人三级黄色视频| 亚洲精品456在线播放app | 精品日产1卡2卡| 免费看a级黄色片| 狂野欧美白嫩少妇大欣赏| 12—13女人毛片做爰片一| 欧美乱妇无乱码| 熟女电影av网| 亚洲成av人片免费观看| 1000部很黄的大片| 给我免费播放毛片高清在线观看| 国产亚洲av高清不卡| 18美女黄网站色大片免费观看| 婷婷亚洲欧美| 母亲3免费完整高清在线观看| 两性午夜刺激爽爽歪歪视频在线观看| а√天堂www在线а√下载| 香蕉久久夜色| 观看美女的网站| 精品久久久久久久人妻蜜臀av| АⅤ资源中文在线天堂| 国产1区2区3区精品| 国产一区二区在线av高清观看| 精品国内亚洲2022精品成人| 久久精品综合一区二区三区| 久久久久久久久中文| 欧美性猛交黑人性爽| 中文字幕高清在线视频| 欧美性猛交╳xxx乱大交人| 国产激情欧美一区二区| 国产精品爽爽va在线观看网站| 中文字幕高清在线视频| 久久中文字幕人妻熟女| 日本熟妇午夜| 欧美黑人巨大hd| 三级男女做爰猛烈吃奶摸视频| 91精品伊人久久大香线蕉| 国产白丝娇喘喷水9色精品| 欧美日本亚洲视频在线播放| 麻豆精品久久久久久蜜桃| 汤姆久久久久久久影院中文字幕 | 干丝袜人妻中文字幕| 村上凉子中文字幕在线| 午夜福利成人在线免费观看| 热99在线观看视频| 国产亚洲av片在线观看秒播厂 | 成人国产麻豆网| 麻豆精品久久久久久蜜桃| 成人高潮视频无遮挡免费网站| 亚洲在线观看片| 看免费成人av毛片| 国产欧美另类精品又又久久亚洲欧美| 亚洲国产精品久久男人天堂| 日韩欧美精品v在线| 国产视频内射| 在线观看av片永久免费下载| 亚洲欧美日韩高清专用| av福利片在线观看| 欧美zozozo另类| 国产伦在线观看视频一区| 热99在线观看视频| 国产乱来视频区| 国产白丝娇喘喷水9色精品| 亚洲内射少妇av| 蜜桃亚洲精品一区二区三区| 欧美日韩一区二区视频在线观看视频在线 | 亚洲国产精品久久男人天堂| 黄片wwwwww| 亚洲成人久久爱视频| 国产精品一区二区在线观看99 | 国产精品人妻久久久影院| 最近中文字幕高清免费大全6| 男女那种视频在线观看| 亚洲精品日韩在线中文字幕| 性插视频无遮挡在线免费观看| 看片在线看免费视频| 人人妻人人澡欧美一区二区| 91在线精品国自产拍蜜月| 国产综合懂色| 日本免费一区二区三区高清不卡| videossex国产| 精品少妇黑人巨大在线播放 | 可以在线观看毛片的网站| 欧美成人午夜免费资源| 全区人妻精品视频| 综合色av麻豆| 国产视频内射| 小说图片视频综合网站| 亚洲国产精品成人综合色| 国产激情偷乱视频一区二区| 日韩大片免费观看网站 | www.av在线官网国产| 狂野欧美激情性xxxx在线观看| 黄色欧美视频在线观看| 七月丁香在线播放| 亚洲av电影不卡..在线观看| 只有这里有精品99| 狂野欧美白嫩少妇大欣赏| 男的添女的下面高潮视频| 国产老妇伦熟女老妇高清| 一个人看的www免费观看视频| 免费大片18禁| 熟妇人妻久久中文字幕3abv| 长腿黑丝高跟| 99热这里只有是精品50| 久久99热这里只有精品18| 99久国产av精品| 黄片无遮挡物在线观看| 狂野欧美白嫩少妇大欣赏| 天堂中文最新版在线下载 | 亚洲,欧美,日韩| 亚洲最大成人av| 一级毛片aaaaaa免费看小| 爱豆传媒免费全集在线观看| 成年免费大片在线观看| av福利片在线观看| 丝袜美腿在线中文| 少妇熟女aⅴ在线视频| 国国产精品蜜臀av免费| 精品久久久久久久久久久久久| 亚洲成人中文字幕在线播放| 亚洲人与动物交配视频| 永久网站在线| 18禁在线无遮挡免费观看视频| 国产伦一二天堂av在线观看| 国产黄a三级三级三级人| 久久人妻av系列| 亚洲精品乱久久久久久| 精品人妻一区二区三区麻豆| 哪个播放器可以免费观看大片| 最近的中文字幕免费完整| 国产一级毛片七仙女欲春2| 亚洲欧美清纯卡通| 九九在线视频观看精品| 18禁在线无遮挡免费观看视频| 人妻制服诱惑在线中文字幕| 青春草亚洲视频在线观看| 国产色婷婷99| 色吧在线观看| av线在线观看网站| 亚洲精品日韩av片在线观看| 精品无人区乱码1区二区| 国产三级中文精品| 在线免费十八禁| 午夜福利在线观看免费完整高清在| 午夜福利在线观看吧| 网址你懂的国产日韩在线| 日本爱情动作片www.在线观看| 国产午夜精品论理片| 久久久精品欧美日韩精品| 人人妻人人看人人澡| 国产亚洲午夜精品一区二区久久 | 国产高清国产精品国产三级 | 亚洲成人av在线免费| 观看美女的网站| 国产伦一二天堂av在线观看| 舔av片在线| 亚洲无线观看免费| 我要搜黄色片| 亚洲成av人片在线播放无| 91久久精品国产一区二区三区| 丝袜喷水一区| 亚洲av免费在线观看| 亚洲av成人精品一区久久| 国产成年人精品一区二区| 国产一区有黄有色的免费视频 | 深夜a级毛片| 亚洲精品456在线播放app| 精华霜和精华液先用哪个| 亚洲美女搞黄在线观看| 国产精品av视频在线免费观看| av线在线观看网站| 热99在线观看视频| 久久久久久久久久久丰满| 九九久久精品国产亚洲av麻豆| 午夜精品国产一区二区电影 | 亚洲最大成人手机在线| av天堂中文字幕网| 亚洲综合精品二区| 网址你懂的国产日韩在线| 毛片一级片免费看久久久久| 欧美最新免费一区二区三区| 男人的好看免费观看在线视频| 人体艺术视频欧美日本| 亚洲va在线va天堂va国产| 国产又色又爽无遮挡免| 亚洲国产欧美人成| 美女被艹到高潮喷水动态| 亚州av有码| 内地一区二区视频在线| 草草在线视频免费看| 别揉我奶头 嗯啊视频| 日本wwww免费看| 一卡2卡三卡四卡精品乱码亚洲| 深夜a级毛片| 嘟嘟电影网在线观看| 精品久久久久久久久亚洲| 精品99又大又爽又粗少妇毛片| 精品午夜福利在线看| 久久精品国产亚洲av涩爱| 亚洲最大成人手机在线| 97人妻精品一区二区三区麻豆| 国产 一区 欧美 日韩| 99九九线精品视频在线观看视频| 在线天堂最新版资源| 国产69精品久久久久777片| 七月丁香在线播放| 国产午夜精品论理片| 国产精华一区二区三区| 成人三级黄色视频| 亚洲在线自拍视频| 国产国拍精品亚洲av在线观看| 又黄又爽又刺激的免费视频.| 日韩欧美三级三区| 亚洲国产精品久久男人天堂| 搡女人真爽免费视频火全软件| 久热久热在线精品观看| 26uuu在线亚洲综合色| 亚洲国产精品国产精品| 国产成人91sexporn| 中文在线观看免费www的网站| 美女国产视频在线观看| 亚洲人成网站高清观看| 国产成人福利小说| 国产精品人妻久久久久久| 亚洲av一区综合| 在线观看66精品国产| 久久久色成人| 91久久精品电影网| 成人性生交大片免费视频hd| 午夜福利高清视频| 午夜激情福利司机影院| 自拍偷自拍亚洲精品老妇| 麻豆av噜噜一区二区三区| av在线观看视频网站免费| 麻豆av噜噜一区二区三区| eeuss影院久久| 亚洲人成网站在线播| 久久久久精品久久久久真实原创| 欧美成人午夜免费资源| 国产精品一及| 国产淫语在线视频| 国产高清国产精品国产三级 | 尾随美女入室| 亚洲aⅴ乱码一区二区在线播放| 一级黄色大片毛片| 国产精华一区二区三区| 欧美区成人在线视频| 欧美xxxx黑人xx丫x性爽| 内射极品少妇av片p| 91久久精品国产一区二区成人| 日本免费在线观看一区| 久99久视频精品免费| 久久99蜜桃精品久久| 九九爱精品视频在线观看| 亚州av有码| 中文资源天堂在线| 一个人观看的视频www高清免费观看| 麻豆久久精品国产亚洲av| 成人鲁丝片一二三区免费| 热99在线观看视频| 成人一区二区视频在线观看| 中文字幕熟女人妻在线| 伊人久久精品亚洲午夜| 丰满乱子伦码专区| 国产成人精品一,二区| 亚洲怡红院男人天堂| 蜜桃亚洲精品一区二区三区| 只有这里有精品99| 亚洲成人久久爱视频| 午夜福利在线观看免费完整高清在| 亚洲精品影视一区二区三区av| 日日摸夜夜添夜夜添av毛片| 国产成人福利小说| 国产高清国产精品国产三级 |