• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Decision-Making in Driver-Automation Shared Control: A Review and Perspectives

    2020-09-02 04:00:12WenshuoWangMemberIEEEXiaoxiangNaDongpuCaoJianweiGongMemberIEEEJunqiangXiYangXingandFeiYueWangFellowIEEE
    IEEE/CAA Journal of Automatica Sinica 2020年5期

    Wenshuo Wang, Member, IEEE, Xiaoxiang Na, Dongpu Cao, Jianwei Gong, Member, IEEE,Junqiang Xi, Yang Xing, and Fei-Yue Wang, Fellow, IEEE

    Abstract—Shared control schemes allow a human driver to work with an automated driving agent in driver-vehicle systems while retaining the driver’s abilities to control. The human driver,as an essential agent in the driver-vehicle shared control systems,should be precisely modeled regarding their cognitive processes,control strategies, and decision-making processes. The interactive strategy design between drivers and automated driving agents brings an excellent challenge for human-centric driver assistance systems due to the inherent characteristics of humans. Many open-ended questions arise, such as what proper role of human drivers should act in a shared control scheme? How to make an intelligent decision capable of balancing the benefits of agents in shared control systems? Due to the advent of these attentions and questions, it is desirable to present a survey on the decision making between human drivers and highly automated vehicles, to understand their architectures, human driver modeling, and interaction strategies under the driver-vehicle shared schemes. Finally, we give a further discussion on the key future challenges and opportunities. They are likely to shape new potential research directions.

    I. Introduction

    INCREASINGLY powerful technologies have facilitated autonomous vehicles to prevent traffic accidents, improve traffic efficiency, and make cars available for everyone, but many social and technical obstacles remain on the road to fully autonomous driving [1], [2]. Overcoming these challenges to enable autonomous cars to drive in highly complex driving situations safely may require some time [3].Analogous to the levels of automation which range from complete human control to complete computer control [4], [5],the Society of Automotive Engineer (SAE) International defines five levels to describe autonomous vehicles and have been adopted by the U.S. National Highway Traffic Safety Administration (NHTSA) in Washington, D.C., as the government’s template. As a transition to autonomous vehicles, partially automated car, in which the human driver and automated driving agent1An automated driving agent refers to as a well-designed automatic controller in automated vehicle systems.share and complete a driving task, becomes a compromise plan before the era of fully autonomous vehicles [6]. Level 3 of automated driving, called conditionally autonomous driving, enables vehicles to mutually transit driving modes between fully automated driving and full manual control [7], but this would degrade the vehicle performance, primarily when transferring the vehicle control authority from car to a driver [8], [9]. Because it requires a transition period for the driver to be resumed in the driving process [10], which can often pose difficulties when the driver has not been actively engaged in the driving process to reacquire situation awareness [7]. Research demonstrates that vehicle automation2Automation is a technology that actively selects data, transforms information, makes decisions, or controls processes [11].harms mental workload and situation awareness [12], [13] and that reaction times increase as the level of automation increases [14]. Another kind of automated driving is called semi-automated driving, in which the automation system does not take full authority from the driver,that is, the driver should keep eyes firmly on the road, though the driver feet off the pedal, hand off the steering wheel [8].However, humans will be bored and distracted during a lowlevel supervision task [8], [15] and will show over-trust [16],[17], neglect [18], and complacency [19] on automated driving systems, which requires a long time to resume control from the automation system in critical situations [10].Therefore, Gordon and Lidberg [8] hold that semiautonomous driving does not alleviate the regular task of anticipating traffic hazards.

    Human interaction with automated driving agents is a kind of human-robot interaction (HRI), and one of the most influential concepts for HRI was supervisory control [20].Supervisory control usually involves a human supervisor setting up the goals while the automated driving agent applying control actions to achieve the goals [21]. However,practically it would be difficult to establish a sharing of control between the human supervisor and the automated agent in this context [22]. Sheridan and Verplank [5] defined two sub-concepts to explain the idea of supervisory control further. One is traded control which requires that both human and computer are active at the same time, and the other one is shared control in which at one time the computer is active, at another the human is [23].

    The shared control scheme presents a tractable paradigm to tackle the driving authority transition issues [24], [25] in Level 3. It features achieving a continuous authority transition between the human driver and automated driving agent.Though the authorized levels of automated driving were issued in 2016, the underlying shared-control concept has already been introduced in other fields [20], [21], [23] and will be the sharp end of cooperation between agents [26].Each agent in the shared control scheme can take their advantages when performing a specific driving task [27]. It is well known that manual control is prone to human errors. On the other hand, fully automated tasks are currently subject to wide-ranging limitations in decision-making and situationawareness. To exploit full potentials of both of human and automation while overcoming the barriers of car-to-driver transition, Mulder et al. [24] presented an entirely different control scheme – shared control systems3The definition of shared control is slightly different over different research fields because there is no single definition for shared control that is used across application domains. More detailed descriptions are referred to see Section II in review paper [21].. The human driver and the automated driving agent continuously share and cooperatively complete a specific driving task, thereby allowing drivers to enjoy driving while keeping in control consistently. Moreover, the shared-control scheme can synergize innate human capacities and technological capabilities to enable us to realize our full potential [28].Previous research experimentally demonstrates that keeping driver’s haptic control authority in the loop with continuous haptic feedback to the driver not only outperforms the conventional binary switches between supervisory and manual control [29] but also reduces distraction on a secondary task[30] and drivers’ workload [31]. Fig.1 presents a drivervehicle system where the human driver and automated driving agents cooperatively share and achieve the same driving task.The shared control over manual control has shown the advantages in many applications such as lane keeping assistance [35], [36] and steering assistance system [31],[37]–[39].

    To some extent, driver-vehicle shared control is a kind of driver assistance system, and from this point of view, which includes three categories: perception enhancement (e.g.,informational assistance), action suggestion (e.g., decision or action selection), and function delegation (e.g., action

    Fig.1. Illustration of the interactions between the human driver and the automated driving agent [32]–[34].

    implementation) [4], [40]–[42]. The first two types have been reviewed in [3], except for the third type, in which both human drivers and automated driving agents can exert their inputs to vehicles simultaneously to carry out a specific task such as the active steering assistance systems. A welldesigned driver-vehicle shared control system should allow all engaged agents to know each other very well [43], which requires addressing the following fundamental research questions:

    1) What kind of role should the human driver act in the shared control system with changing situations?

    2) How to allocate the driving responsibility and authority according to the ability of two agents?

    3) How to on-line evaluate the respective trust levels among drivers and automated driving agents?

    4) What are the temporal scales of human adaptation and learning in changing situations?

    5) What novel system identification techniques for driver state and intent exist that could allow us to study time-varying and possibly nonlinear shared control systems?

    Although many works of literature have been done for specific topics, there is no paper to review and discuss these research questions comprehensively. To bridge the gap, we provide an overview of the field of decision-making scheme design and human driver modeling in shared control systems,by reviewing more than 200 closely related literature covering the keywords: shared control, driver model, shared cognitive control, self-driving, and human-automation/robot/computer interaction. Instead of reviewing rigorous mathematical algorithms of decision-making and controller design, we mainly focus on the scheme design of decision-making,human driver modeling, and the open issues with potential solutions in shared control systems, thus benefiting researchers working on the considered topic. Some other partially related literature on psychology and ergonomics is only involved and referred without in-depth discussion because of page limitation. Section II describes the underlying architectures of driver-vehicle shared control systems.Section III reviews the decision-making of two agents in the driver-vehicle shared control system from the state-of-the-art literature. Section IV details human driver modeling.Section V shows and discusses some open-ended, challenging,inevitable scientific questions. Section VI gives further discussion and conclusion.

    II. Shared Control Architectures

    Before giving an insight review, we first discuss the architecture of the driver-vehicle shared systems. A sharedcontrol system consisting of a human driver and an automated driving agent refers to as a two-agent system that is capable of accepting and executing commands from a human driver, or an automated driving agent, or a combination of the two [44].According to the role that the human driver plays in the driver-vehicle shared systems, the shared control can occur at two different levels: the task level and the servo level (as shown in Fig.2).

    Fig.2. Two kinds of shared control architectures: (a) Task-level shared control; and (b) Servo-level shared control. rh and rc represent the expected trajectories to be followed for human driver and automated driving agent,respectively; us is the input of vehicle, uh and uc are the human driver’s o peration output and automated driving agent’s output, respectively.

    A. Task-Level Shared Control

    In the task-level shared control4The task-level shared control refers to as a shared-control system that allows human drivers to decompose a whole driving task into subtasks and allocate some of them to the automated driving agent (namely, share at the task level[44] rather than the servo level), which is slightly different from the traded control [5], [20], [21].scheme [44]–[46], human drivers usually act as a guide and deliver a task-specific command to the automated driving agent. Namely, the human driver can allocate subtasks to the automated driving agent while authorizing other subtasks to achieve a complete driving task. After being informed the subtasks, the automated driving agent will perform the subtasks based on the current situation condition and predefined algorithms to cooperatively achieve the whole driving task together with the human driver. The task-level shared control scheme fully exploits the strength in both machines and humans, which could alleviate the driving burden of a human. A very intuitive example is that human drivers manually activate the adaptive cruise control (ACC)systems when driving on the highway [47] to allocate longitudinal control authority to the ACC agent while authorizing lateral control by himself/herself.

    B. Servo-Level Shared Control

    Servo-level control in the driver-vehicle system usually focuses on servo control. The control input ( us) to a vehicle is typically the combination of human drivers’ operations ( uh)and automated driving agents’ output ( uc), as shown in Fig.2(b).In the servo-level shared control, differing from the task-level shared control situations where the automated driving agent will take over the task-specific control, human drivers will always be engaged in the control process of vehicle movement at the servo level. The combination of outputs from human drivers and automated driving agents should be well designed.Analogous to human-robot shared control, researchers[48]–[52] combine them intuitively by

    where λ ∈[0,1] is the weighted coefficient to adjust the proportion of uhand ucin us. The allocation of driving authority for a human driver and an automated driving agent is determined by λ which can be either fixed or continuously adaptive. Pure human control is achieved when λ=1 and pure automatic control when λ=0. Therefore, the value of λ can impact the shared control scheme by determining if the driver is presented in the loop. This parameter can be modulated manually (e.g., in [53]) and automatically based on the driver’s states in certain situations in which the driver needs assistance to perform more rapidly and safely (e.g., in [49]). The design of the λ will be discussed in Section III.

    Different from linearly combining the human driver’s input with the automated agent’s input by (1), some probabilistic models [54] have also been proposed by explicitly taking into account the uncertainty in the interaction and modeling this combination as a joint probability distribution [55].

    The approaches of combining uhand uccan differ from each other in different servo systems. As a consequence, the servo-level shared control can be further divided into direct shared control and indirect shared control [56], [57].

    1) Direct Shared Control: The direct shared control allows both human drivers and automated driving agents to simultaneously exert actions on a control interface, of which the output remains the direct input to vehicle systems, as shown in Fig.2(b). Such systems are usually haptic shared control [58] since both human drivers and automated driving agents will directly influence the inputs on the haptic control interface (e.g., steering handwheel and brake/throttle pedals).Also, human drivers can even percept the assistance torque applied by automated driving agents through the shared haptic interface. A general architecture of haptic shared control can be seen in [59].

    2) Indirect Shared Control: Differing from the direct shared-control scheme, the indirect shared control scheme shapes the input to the controlled vehicle system by mixing the out of control interface (usually as a result of human contributions) and output of the automated driving agent [37],[38], [58], formulated by us=g(rh,rc,uh) as shown in Fig.3.A typical application in the human-vehicle systems is steerby-wire (SBW) systems [60] which estimates the driver’s desired steering angle from the driver operations and then generates and applies the steering angle directly to front

    Fig.3. Illustration of an indirect shared control scheme [56], also called“input-mixing shared control” [58].

    wheels.

    III. Decision-Making in the Servo-Level Shared Control Scheme

    This paper mainly focuses on the decision-making in servolevel shared control scheme rather than the task-level shared control scheme. A well-described model of the entire drivervehicle system can offer one a better understanding of how each subsystem works and what the relationship between them is. Some researchers hold that driver modeling should be integrated into the road-vehicle system for control purposes to improve the mutual understanding between the driver and automation [61], [62]. Hence, a driver-in-the-loop (DHL)5The “l(fā)oop” can refer to an information processing control loop (i.e., attentive to driving task) or a sensory-motor control loop (i.e., vehicle control), or both[63]. Here, the “l(fā)oop” refers to as a sensory-motor control loop.vehicle model [61], [64], [65] is usually incorporated into a shared control system and then describes the complete DIL systems from a control perspective.

    For this purpose, first, the road-vehicle dynamic model is usually formulated using the state-space representation

    where x is the road-vehicle state, A is the road-vehicle system matrix, B is the input matrix from road-vehicle system, Bwis the system disturbance matrix, ucis the controller input, uhis the human operation input, w is a system disturbance. Then the human driver is formulated from a control perspective as

    where rhis the driver’s desired trajectory or reference trajectory. H(rh,x) represents human driver model and can be formulated from control perspective [66], stochastic perspective [67]–[69], or cognitive perspective [70], which will be discussed in Section IV. The driver model towards the applications to steering system control (or lateral control,path-following control) is the preview driver model and its extensions (see review article [66]) since it is easy to integrate them into the state-space-based vehicle model. Human drivers output their operations and apply to the vehicle by comparing the desired trajectory and the current trajectory through their internal model [71], [72]. The internal model can estimate the current/upcoming states of the subject vehicle and trajectories of surrounding objects (e.g., vehicles, pedestrians, and bicycle users). Substituting (3) into (2), the driver-vehicle model (i.e.,DIL model) can be formulated by a new state-space representation

    where z is the augmented state consisting of human driver model state and road-vehicle system state, ADILis the drivervehicle system matrix, Bhis the human input matrix, and Bcis the controller input matrix. The common DIL models based on the state-space representation can be found in [61], [64],[73]–[77] and some popular cases are also listed in Table I.The state-space DIL model provides an analytical way to assess the stability of the shared control system [61], [90] and an standard way to design controllers [50], [91]. A welldesigned decision-maker and controller for the shared control system should assist human drivers driving safely and smoothly while without causing any conflicts with human drivers. Based on the incorporated DIL model (4), given a desired trajectory/reference and the human driver operations,the optimal controller input can be obtained by solving an optimal problem in the general form

    where J(·) is the objective function that could encompass human driver’s input and other constraints (e.g., vehicle dynamics and human driver’s physical limitations). To solve the optimization problem between a human driver and an automatic controller, one of the biggest challenges is to formulate the relationship between uc, uh, and us, e.g.,allocation of the control authority, λ. In what follows, the shared control strategies can be formulated according to prior knowledge or dynamic programming. Thus, two ways to design the decision-making strategies are listed: rule-based and game theory-based.

    A. Rule-Based

    For the rule-based method, one direct way is to use (1) with the requirement of designing λ(t). Most research predefined a different kind of rule according to prior knowledge, which can be roughly grouped into three categories and discussed as follows.

    1) Piecewise Function: Due to the complexity of dynamic environment and disturbance, an intuitive way to design λ(t) is using rule-based piecewise function. The piecewise function is primarily developed to tackle the shared-control issues in robotics such as wheelchair and industrial robots [92], [93]and then is introduced to intelligent vehicles afterward because of its robustness and practicality in terms of controller design [94]–[96]. For instance, Jiang and Astolfi [92], [96]defined three space sets to divide the reachable set into three parts by judging the level of safety — safe, close, and dangerous. Correspondingly, a three-level piecewise function consisting of the safe Rs, hysteresis Rh, and dangerous subsets Rdwas proposed to design λ(t):

    If human behaves “dangerously” then the feedback controller (i.e., automated agent) is active and resumes the control authority; if human behaves “safely” then the vehicle only responses to human’s operations; if human behaves in

    “hysteresis” set, then the vehicle runs under a predesigned shared-control law.

    TABLE I DIL Model Using State-Space Representation Towards the Applications in Vehicle Dynamic Control

    Besides, the piecewise function is an analytical way to take account human factors and driving situations in the light of its practical integration with human’s prior knowledge. For example, Li et al. [49] utilized two piecewise functions to assess driving situations and vehicle performance separately based on their empirical knowledge and then fused these piecewise functions as the λ(t). Saito et al. [25] used an exponential function of the steer-wheel torque applied by the human driver to estimate how much torque assistance the assisting agent should deliver.

    2) Exponential Function: The second approach to obtain a seamless and smooth λ is to use the family of the exponential function:

    where ζ(t) is the activity factor, α0and α1are the tuning parameters, and αsafeis the parameter to guarantee safe and model convergence. The exponential function has been widely used in human-robot shared control, and then introduced to tackle control authority transfer and allocation in humanvehicle shared control afterward. For example, Sentouh et al.[50] integrated the discrete driver state (distraction) into the exponential function to obtain a continuous shared control factor λ(t). Wang et al. [37] designed a shared steering control law using an exponential function with considering different driving styles to improve vehicle performance and reduce drivers’ workload when taking curve negotiation. However, a religious setting should be made when applying this exponential function, since the derivative of (7) could be discontinuous when the denominator switches between positive and negative.

    In order to determine a continuous weighted coefficient ( λ)for the shared control systems, researchers also combine the piecewise function and exponential function to evaluate the safety level for decision-making in human-robot shared control [97]. This kind of shared control strategy has been carefully borrowed to improve human-vehicle shared control performance. For example, Sentouh et al. [50] utilized the function in the format of (7) to obtain a continuous authority allocation factor from a discrete driver drowsiness monitoring factor. Besides, some probabilistic shared control strategies towards complex, dynamic environments were also proposed by modeling both the human’s intentions and the automated agent as a probabilistic function to improve the shared control performance with a exponential function [91], [98].

    3) U-Shape Function: Another approach is to directly compute how much assistance the driver needs based on the U-shape function, i.e., the relationship between drivers’workload and performance as well as the need for assistance,as shown in Fig.4. For instance, Nguyen et al. [64], [101]designed an assistance torque Ta=μ(a)Tsto reduce drivers’workload and improve vehicle performance using the U-shape function of drivers’ activity μ(a), where Tsis the required input torque from the vehicle, and a is the drivers’ activity denoted as steering angle. Oufroukh and Mammar [88] also proposed a similar computation model to compute the assistance torque using U-shape representations during lane keeping or obstacle avoidance maneuver.

    B. Game-Theory-Based

    Differing from the rule-based method where a predefined shared control law λ(t) is adopted, some researchers treated human drivers and automated driving agents in the shared-

    Fig.4. U-shape representation of driver’s need for assistance [8], [99],[100].

    control system as two-player with dynamic interaction, as shown in Fig.5. The two-player assumption enables gametheoretic approaches to be available to tackle the relations between human drivers and automated driving agents.

    Fig.5. Dynamic games between a human driver and the automated driving agent.

    Game-theoretic approaches have been widely applied to tackle the dynamical decision-making problem where two or more agents make decisions that influence one another welfare [102] such as vehicle-to-vehicle [103]–[105], grid-tovehicle [106], [107], collision avoidance at an intersection[108]. Applications of game theory in modeling road user behaviors and traffic or transportation can refer to the review literature [109], [110]. Na and Cole made a comprehensive classification for dynamic games as shown in Fig.6.Depending on the interactive type between human drivers and automated agents, the dynamic games between two agents can be classified into noncooperative and cooperative games[112], as illustrated below.

    1) Noncooperative: Drivers and automated driving agents consider themselves as individuals and concentrate on pursuing their interest. More specifically, in the noncooperative game theory, the strategy type of driver and automated driving agent can be derived using Nash equilibrium and Stackelberg equilibrium. A Nash equilibrium emerges in situations, where drivers and automated driving agents derive their strategies by considering others’ strategies,and they act simultaneously. A Stackelberg equilibrium emerges in situations, where one agent (i.e., human driver or automated driving agent) is the leader, and the other one serves as a follower.

    Fig.6. Classification of dynamic games [111].

    2) Cooperative: Drivers and automated driving agents have a sense of collectivity and attempt to enter into a binding agreement of interest, where the goal of each agent is identical, and their strategies are derived from global optimality [113].

    Usually, the two agents are assumed to be rational with individual objectives [111], [114]–[116]. The vehicle is controlled by a human driver and an automated driving agent,modeled as

    The goal of a human driver and automated driving agent is to minimize their objective function.

    1) Noncooperative Nash Scheme, where the automated

    driving agent will compensate for human drivers’ erroneous actions to imitate the possible opposite effects, and simultaneously, human drivers will also estimate action applied by the automated driving agent, and to neutralize its influence by further changing her/his inputs;

    TABLE II Game-Theoretical Applications in the Shared Control Between Human Driver and Automated Driving Agent

    2) Noncooperative Stackelberg Scheme, where the automated driving agent will compensate for human drivers’erroneous actions to imitate the possible opposite affects, and sequentially human drivers will generate actions when they have a very well knowledge of the actions applied by automated agents, and vice versa;

    3) Cooperative Pareto Scheme, where both human drivers and automated driving agents try to account for each other’s desired trajectory, and simultaneously will react to both each other’s actions.

    The Nash and Stackelberg equilibriums can be employed to solve the closely coupled optimization problem in drivervehicle interactions. Either analytical or approximated solutions can be derived as the human driver’s and the automated driving agents’ control actions. When the approximated solution is concerned, the following two expressions hold:

    For example, Flad et al. [116] introduced an approximated Stackelberg solution [122] to solve the problem between a human driver and ADAS controller (i.e., automated driving agent) by treating one of them as a leader and the other one as a follower. In [77], Li et al. designed a continuous role adaptation of human-robot shared control using the Nashequilibrium, where human and the automatic controller (i.e.,automated driving agent) can simultaneously exert control to the robot, instead of directly using the Stackelberg solution[114], [107]. An adaption law was also designed by comparing the difference between measured human drivers’inputs and predefined Nash equilibrium. Besides, the game theory could also integrate the uncertainty from human drivers and external factors into the shared control system [121] with a stochastic dynamic programming solver.

    Although the above mentioned two typical methods offer us ways to design the shared control strategies between a human driver and automated driving agents, the two agents could also fail to cooperate when a wrong estimation of current states regarding driving situation perception and human driver intent occurs. Also, a poor-designed shared control strategy could even bring four main adverse effects [123], [124]: loss of expertise, complacency, trust, and loss of adaptivity.Therefore, studies dealing with human-vehicle cooperation have to consider the human’s characteristics such as uncertainty according to reduce the conflicts with the driver.In what follows, we will discuss about modeling human drivers from different perspectives.

    IV. Human Driver Modeling

    In order to obtain well-designed shared control interactions between a human driver and a highly automated vehicle,Norman [125] stated that human must always be in control,must be actively engaged and adequately informed, and that human and automated vehicle must understand each other’s intents correctly in the complex driver-vehicle systems.Therefore, understanding and modeling human drivers’sensory dynamics, cognition processes, hidden states, and operation characteristics are equally important as dynamic vehicle systems for driver-vehicle shared control systems.Human drivers usually complete a driving task at three levels[126], [127] (Fig.7): strategic level, tactical level, and operational (control) level. The first two levels involve cognition while the third level involves execution. According to the previous review of all literature, we introduce and discuss the driver model from its functional modules,modeling approaches, and driver’s intent inference as well as state detection.

    Fig.7. A hierarchical structure for modeling human drivers when driving[126]. EI = Environment inputs.

    A. Functional Module

    1) Sensory Dynamics: A recent review in [128]demonstrates that drivers’ sensory dynamic characteristics play an important role in driver-vehicle system design,making them interact more friendly and safe. Driver’s sensory dynamics used for vehicle speed and direction control mainly include [128] (ranked by the level of importance for a regular driving task [71]):

    a) Visual — The visual system is the human driver’s only means of detecting the future road trajectory. A general review of overall literature demonstrates that visual information is the highest significant in normal driving process, accounted for about 90% of all sensory information[71]. Driver’s visual information (e.g., eye gaze [129], [130])can also reflect driver’s underlying intentions (e.g., lane change intents [131], readiness to take over from automation[132]), mental or physical states (e.g., fatigue detection [133]),and upcoming actions [134].

    b) Vestibular and kinesthetic — Human drivers usually use the motion information (e.g., acceleration and rotation of vehicles) derived from vestibular and kinesthetic channels as a supplement to visual information for control task, which could also contribute to the human combined arm-trunk motion[135].

    c) Somatosensory — The somatosensory information used in control of vehicle steering and speed mainly includes tactile and haptic information such as pressure on the gas/brake pedal[136] and steering torque on the steering handwheel [39].

    d) Auditory — The auditory information usually is used as a supplementary cue within a multi-channel environment in a normal driving process.

    The sensory dynamics have their physical characteristics such as time delay, perception limitations, and coherence zones because of human’s physical, biochemical, mental limitations and the human’s ability to perceive and process information. A driver model which integrates human drivers’sensory dynamics could help understand humans, thus for improving comfort, safety and driveability of driver-vehicle shared control systems and reducing the incompatibility6Compatibility here is referred to as the quality describing the fit or match[32] between a human driver and automated driving agents, regarding the outer (e.g., interfaces) and inner (e.g., cognition) interactions [137] between them.or negative interference [123] between the human driver and automation. The haptic information on the steering wheel has been widely used in the shared steering control systems to reduce the driver’s mental/physical workload and to improve driving skills [29], [138], [139]. A sensorimotor driver model was developed to improve the shared-control performance by considering both visual and kinesthetic perception and including compensatory and anticipatory processes [81].

    2) Cognition: John A. Michon has mentioned the cognitive driver model from a critical view of introducing the behavioral sciences and psychology to understand the human decision-making processes [126]. Much of the decisionmaking generated by human drivers is over discrete actions,such as choosing whether/when/how to lane change [140]. To capture these discrete features, researchers modeled human drivers’ high-level tactical behavior (e.g., speed selection and decision making) and strategic behavior (e.g., route planning and navigation). One of the most utilized means is based on the “adaptive control of thought-rational (ACT-R) [141]”cognitive where the discrete nature of drivers’ control actions is captured from a cognitive perspective. For example,Salvucci et al. developed an integrated cognitive pathfollowing driver model [142] and lane-change driver model[143] using the combination of the ACT-R cognitive architecture and perceptual-motor process.

    Some researchers also developed cognitive driver models based on on-hand knowledge or with new insights in experimental data. Misener et al. [144] developed a cognitive car-following model to avoid rear-end crashes with a stopped lead vehicle by fusing current knowledge derived from experimental data. Liang et al. [145] developed a system to detect driver distraction in real-time by cognitively analyzing three indicators: how to define distraction, which data were input to the model, and how the input data were summarized.Liang also demonstrated that combining cognitive and visual distractions can improve vehicle performance than each of them [146]. Much more literature on modeling and analyzing drivers’ cognitive distraction can see references [147]–[150].A review paper for modeling driver behaviors in a cognitive architecture refers to [70], [72].

    Based on the cognitive driver model, a cognitive assist can be potentially provided in the context of driver-vehicle shared control to reduce harmful interference between two agents.Cai and Lin [151] proposed a coordinating cognitive assist to determine when an assist should be provided and how much assistance to be supported for a steering assist control system.The cognitive assistance was divided into three stages to assist drivers in acquiring information, analyzing information and making a decision, and implementing action [4].

    3) Neuromuscular-Skeletal Dynamics: Existed research has demonstrated that a good understanding of neuromuscularskeletal dynamics has constraints upon the dynamics of perception-action coupling [152] and is significantly vital for a well-designed shared control system, for example, avoiding subtle conflict between a human driver and automated agent[153]. For human drivers, the neuromuscular-skeletal dynamics in human-vehicle systems mainly include the arm and foot, representing the lateral and longitudinal control,respectively.

    For the dynamic properties of a driver’s holding a steering wheel, Pick and Cole [154], [155] investigated the effect of the driver opposing a constant offset torque and the effect of the driver co-contracting the muscles, and found that both actions will increase the stiffness and damping of the arms. A linear model of the neuromuscular system, muscles, limbs,and the vehicle was then set up and applied to a driver simulator [156], a guideline to shared control [58], a pathfollowing driver–vehicle model [157], and a driver lateral control model [80], [158]. Besides, the characteristics of driver neuromuscular dynamics are different from each other and affected by steering systems (e.g., active and passive) and hand positions on the steering wheel [159], which should be considered when designing a controller.

    For the dynamic properties of a driver’s low leg and foot hitting the gas or brake pedal, researchers mainly focus on the ankle-foot neuromuscular dynamics, which enables us to gain insight into both responses to visual feedback and haptic feedback. Abbink et al. [136] investigated the biomechanical properties of the ankle-foot complex (i.e., admittance) [160]and developed a driver support system that uses continuous haptic feedback on the gas pedal to inform drivers of the separation to the lead vehicle.

    B. Modeling Approaches

    1) Control-Theoretic Driver Model: Understanding the potential and the limits of a tightly coupled driver-vehicleroad system are not trivial. Many kinds of driver models towards the application in vehicle dynamics control can be found in review articles [66], [161], [162], in which literature before 2014 was listed, including lateral and longitudinal driver models. Table I lists some popular driver models toward vehicle dynamics control applications and indicates that the single/two-point visual preview driver models or their extensions are highly preferred since they are easy to be integrated with vehicle models. However, one of the most significant limitations is that the single/two-point visual preview driver models assume that human drivers’ references or desired trajectories are exactly known, which is not always available in the real application. Besides, for such two-point visual preview models, drivers’ essential physical characteristics such as the reaction time delay, anticipatory and compensatory can be reflected, but not cognitive characteristics such as neuromuscular dynamics of driver arms and decision-making processes. For example, the controltheoretic driver models (see review [66]) usually ignore issues of whether or how to perceive the model inputs from the external environment and how drivers correctly interact with other automatic controllers [70] via visual, haptic, or auditory sensors [71].

    Most control-theoretic driver models have been used in the driver-vehicle shared control systems with a game-theoretic scheme. With this purpose, the human driver and the automated driving agent should be first modeled, which makes it feasible to estimate each other’s actions. One of the most popular approaches for modeling automated driving agents is using control theory (e.g., model predictive control)since it can describe drivers’ ability to predict vehicle’s future states [163] based on their internal model or individual’s driving skills [164]. Na and Cole [111], [114] applied the combination of the distributed MPC and linear quadratic dynamic optimization (LQDO) to formulate human drivers and automated driving agents. The objective function with quadratic structure is also widely used to design an automated driving agent considering human drivers’ forthcoming motion primitives [115] or drivers’ haptic inputs on the control interface [77].

    2) Learning-Based Driver Model: Though the driver models as mentioned above could describe and predict drivers’behavior, actions, and states with relatively satisfied performance, they did not concern the dynamic, stochastic decision-making processes of driver behavior, which requires models capable of connecting temporal and spatial processes.For this purpose, some researchers also utilized learning-based approaches to deal with highly nonlinear properties of driver behaviors, such as neural networks [165], [166]. Research in[68], [167]–[169] demonstrates that the Markov models combined with Gaussian mixture models achieve a satisfied performance of capturing driver intent and action. Besides, the Bayesian inferences [170], [171], autoregressive exogenous(ARX) [172], and deep learning [173] were also developed.These learning-based approaches highly depend on the collected training data, and some of them are data-hungry such as deep neural networks [174].

    C. Intent Inference and State Detection

    Correctly inferring drivers’ intents and states is profoundly essential to design an automatic controller capable of delivering an adequate input not only to follow/track the desired trajectories but also to avoid intrusive interventions between the human driver and the automated driving agent[175].

    1) Intent Inference: Steering wheel, as a direct interface,allows human drivers and automated driving agents to act and exchange information in a simultaneous and continuous way[176], [177]. Therefore, human drivers’ intents can be directly captured through the torque applied on the steering wheel by drivers. For example, in a torque-based steering assistance system [39], the automated driving agent uses sensors to obtain the steering torque forced by drivers, and inversely,human drivers can also react to his/her haptic perception information from the steering wheel. Nguyen et al. [74], [101]utilized the torque applied to the steering wheel by drivers as an indicator to compute how much assistance torque an automated driving agent should provide. Li et al. [77]proposed a continuous adaption law for the human-robot shared control system to determine the automated driving agent’s role (i.e., leader or follower) by comparing the measured torque applied by the human to the predefined Nash equilibrium computed through optimal control techniques.

    In addition to using the torque applied on the steering wheel by a human, the dynamic neuromuscular analysis of driver arms or legs can also provide a guideline for shared control design of a steering system [39], [58], [80], [157], [177] and the gas/brake pedal control [136], [177]. For instance, in order to reduce the intrusive intervention between human drivers and automated systems, Ziya et al. [39] modeled human drivers’ steering behavior by combining the neuromuscular response of drivers and the desired steering angle that was a function of vehicle states and road geometry. Moreover, the impedance of a haptic torque was used as an indicator of the drivers’ intents. Some researchers also designed a guidance torque to assist drivers to keep the vehicle in the driving lane[178] and to improve vehicle safety for fatigue-related driver behavior [179].

    Much literature on modeling human drivers’ neuromuscular dynamics of steering behavior has been found in [76], [155],[157]. The model of driver’s arm and steering dynamics usually combines with a path-following control model,obtaining a linear driver model with neuromuscular dynamics[80]:

    where J?are the inertia, damping and stiffness of ?, with ?being driver arm or steering systems; nsis the steering gear ratio; MTis the torque arising from the lateral forces and selfaligning moment; Tmis the muscle torque; and θsis the steering wheel angle. Model (13) has been used to infer drivers’ intents. For example, the changes of damping and stiffness of driver’s arms can reflect whether the automatic controller’s outputs satisfy drivers’ desired trajectory or the conflict level between drivers and automated driving agents[58], [59], [154]. Instead of using indirect cues (e.g., steering angle, vehicle dynamics), direct human-observing cues (e.g.,body gesture, head, hands, feet, and gaze direction) can also be used to predict driver intent [180]–[183].

    In addition to directly using the haptic information, driver intent can also be inferred and predicted according to dynamic driving environments such as peripheral vehicles [184] and vehicle position to lane edges [185]. The dynamic Bayesian networks (DBN) [186], [187], Markov decision processes(MDP) [167], [181], and partially observable Markov decision processes (POMDP) [188] are one obvious place to start,assuming that one can extract the underlying latent processes of driver behavior which is a dynamic and stochastic process.Besides, the correlation between human driver gaze behavior and steering moments will decrease while increasing the allocation ratio of control authority for intelligent driving agents [189], indicating that driver gaze behavior could be used to infer driver’s intent and avoid conflict.

    2) State Detection: Correctly detecting driver state (e.g.,sleepiness, drowsiness, fatigue, distraction, impairment) offers an opportunity to make a practical decision of authority allocation, thus improving vehicle safety. For example,Saito et al. [36] proposed a dual control scheme for lanekeeping assistance systems by detecting the driver’s sleepiness level using eye blinking frequency and facial information. More driver intent detection and inference using direct human-observing cues can refer to the review paper[133], [190].

    Visual distraction or cognitive distraction have been investigated by combining vehicle state [36], [147], [148],drivers’ visual state [145]–[147], [149], and operations [147],[149], [191]. Answers to the question of how to measure a driver’s cognitive distraction have been given in [150].Learning-based approaches such as deep sparse autoencoders[192], deep belief networks or DBNs [188], support vector machines (SVM) [145], [193] have been widely used to detect and classify driver distraction.

    V. Future Challenges and Opportunities

    Due to the limited ability to understand human drivers,many open-ended questions still exist in driver-vehicle shared control systems. This section will present and discuss some open-ended, challenging, inevitable questions regarding the shared control strategies, the trust or over-trust, and authority allocation, followed by future opportunities.

    1) How to design the adaption law or adaptive/adaptable shared control? Also, what kind of role of the human driver should be in driver-vehicle shared control systems?

    In Section III, we have discussed different ways to design adaptive shared control between the human driver and an automated driving agent. Most of them were from U-shape,noncooperative game theory, and the torque applied on the steering wheel by the human. The role of human drivers in highly automated vehicles can be defined as

    a) Convertible role between leader and follower [115](game theory);

    b) Symbiotic relations with the automated driving agent(page 24 in [194]);

    c) Being an active driver, passengers, or passive drivers, but they may still be required to take over control [190];

    d) Or being parallel [46], [195].

    The different roles of human drivers in driver-vehicle systems result in various shared-control paradigms. In terms of the methods researchers utilized, the shared control performance can be improved by considering individual characteristics, classifying the levels of human-automation interaction, and integrating with cognitive psychology.

    The U-shape control law only qualitatively describes the relations between drivers’ workload and the needs for assistance as well as the driving performance, but not a quantitative expression. As a result, the adaption control laws derived from U-shape was greatly different, for example, in literature [64] and [88]. Many factors could cause the difference, such as the diversity in individuals’ driving experience and physical/psychological status. Classifying the type of human driver [196], regarding their abilities and characteristics, and then designing a personalized driver model [197] capable of describing and adapting this driver’s characteristics [198] could be an efficient way to tackle this kind of problems.

    In terms of the game-theory-based adaptive law, researchers usually modeled human drivers by assuming that drivers had a perfect internal model [72] for understanding and predicting vehicle states as the same with an automatic controller, that is,both agents had an identical, deterministic objective functions[114], [115], [119]. In the real world, however, humans driving is not always a deterministic process [199], but in nature, a stochastic and dynamic process [167], [200], and even impaired behaviors (e.g., fatigue and drunk). Therefore,the stochastic driver behavior and divergences among drivers should be modeled and accounted in future work. Besides, the underlying relations (the role of a human driver in the drivervehicle system with shared control schemes) between human drivers and automated driving agents remain open-end.

    In response to the problems above, classifying the types and levels of human-automation interaction [4], [12], [201] and increase the agent’s adaptability [202] could be an applicable approach. The automation functions usually cover four types:information acquisition, information analysis, decision and action selection, and action implementation. Within each of these types, the automation is defined and treated as a continuous level from low (i.e., fully manual) to high (i.e.,fully automatic). Human performance consequences in terms of types and levels constitute primary evaluative criteria for automation design. This approach has been used in driver assistance system design [151].

    Besides, Beetz et al. [203] and Heide and Henning [204]proposed the idea of cognitive car separately — a technological cognitive system that can perceive itself and its environment, as well as collect and structure information in an autonomous way. In the cognitive cars, some key issues remain regarding action implementation (i.e., function delegation):

    a) What kind of actions should be implemented?

    b) When to add the actions?

    c) How to implement the function delegation appropriately?

    To date, we are not able to answer the three questions systematically, but most research focuses on i) the effects of haptic support systems on driver performance; ii) vehicle stability control for collision avoidance, and iii) active steering system with adaptive assisted systems. One of the potentially useful approaches is to consider and built cognitive cooperation based on cognitive psychology, as mentioned in[123]. Li et al. also holds that the cognitive cars will be a new frontier advanced driver assistance system for research [205].Besides, the Petri net modeling has been demonstrated as an alternative solution [206].

    2) What is the appropriate trust? Alternatively, would drivers take a seat as passive occupants, who fully trust their vehicles?

    Over-reliance on automated driving and driver complacency are often problematic [19], thus resulting in, for example, a long reaction time [10]. The driver, who fully and deeply trusts the ability of driving automation, therefore failed to intervene and take manual control even as the driver crashed the car [11], [17]. Fortunately, human drivers have an appropriate level of trust in the automated driving agent, such as being aware that the automation system can better sense and faster response [207] and displaying system situation awareness [208], but sometimes can also make mistakes.Driver performance can also be improved by effectively conveying the limitations of automated driving to human drivers [209], [210]. Besides, an appropriate, elaborate practice for drivers could mitigate the negative impact of over-trust in the automated driving agent on reaction time[41].

    Besides, when the desired trajectories derived from the driver and the automated driving agent are similar, the vehicle inputs from both of them locally differ but combine without conflict. However, when the desired trajectories of two agents are different, things will become intractable and bring a question: which one input should be trusted and exerted to the vehicle? One potential way for tackling this issue is to develop a psychologically-grounded cognitive-physical model [194]capable of correctly describing and predicting the driver’s desired trajectory from the operational level, tactical level, and strategic level by understanding information processes(neuroscience) and cognitive abilities (psychology) of the driver. The cybernetic driver models integrating visual(including anticipation and compensation) systems with neuromuscular systems (or motor processes) are an efficient way to benefit shared control. The readers can refer to the review paper [211] for details about the model structure and parameter identification. Another potential way is to design a metric to evaluate and analyze the conflicts between a human driver and automated driving agent [212]. Many indicators have been developed to evaluate shared control performance,and they mainly concern four aspects: accuracy (e.g., pathtracking errors) [49], [50], [211], safety (e.g., risk metrics such as time to lane crossing) [49], [120], compatibility (e.g.,existing conflicts or not) [50], [211], and robustness (e.g., the vibration in the resonance response of vehicle) [50], [61]. In the haptic shared control system, the directions and periods of the torque applied to the steering wheel by a human driver and automated driving agent are most commonly used. For example, researchers in [211] considered four aspects of steering torque to evaluate conflicts in haptic shared control:consistency rate, resistance rate, contradiction rate, and contradiction level.

    3) Which way is the best transition in authority, ability,responsibility, and control (A2RC)?

    Before presenting the open-ended questions, we should define some basic concepts in interactions between human driver and automated driving agents according to [213]:

    a) Authority, which can be defined by what the human driver or automated driving agent is allowed to do or not to do. Further, the authority can be partially and continuously transferred between human drivers and automated agents.

    b) Ability, which can be defined as the possession of the means or skill to perceive and select adequate action and act appropriately.

    c) Responsibility, which can be assigned beforehand to motivate certain actions and evaluated afterward, where the human driver or automated driving agent is held accountable or to blame for a state or action of the driver-vehicle system and consequences resulting thereof.

    d) Control, which means having the power to influence the vehicle states.

    Fig.8 shows the relations between the four concepts. Based on these basic definitions, the issues that existed in humanautomation interaction systems are also encountered in the human-vehicle shared control systems since the highly automated vehicle is a case-specific automation system [4],[213], for example:

    i) How do we balance between exploiting increasingly powerful technologies and retaining authority for human drivers?

    ii) How can we define clear, safe, efficient and enjoyable roles between human drivers and automated agents?

    Fig.8. Relations between ability, authority, control, and responsibility[213].

    iii) Which of the subsystems of future human-vehicle systems should have which ability, which authority and which responsibility? Alternatively, which system functions should be automated and to what extent?

    iv) What other concepts besides authority and responsibility do we need to describe and shape a dynamic but stable balance between human drivers and automated driving agent?

    According to these predefined terms, there is an allocation problem in terms of A2RC between human drivers and automated driving agents. Most literature only focuses on the control authority between human drivers and automated driving agents, but ignore the relations regarding the A2RC.In the real world, the authority, ability, responsibility, and control are not independent. In [99], an A2RC diagram was developed to analyze and design human-machine systems,which can be used to tackle the transition issues of A2RC in driver-vehicle shared control systems. Besides, Acarman et al.[214] also proposed a control authority transition system for collision and accident avoidance considering drivers’ physical and cognitive capacities as well as the situation/danger/hazard analysis. In addition to the A2RC, the H-metaphor (H-mode or horse metaphor) [215], i.e., a metaphor of the relationships between a horseback rider and the horse, could be as a guide to open new horizons in the shared-control systems, which has shown its remarkable achievement in aircraft co-pilot design[216] and highly automated driving [217].

    On the other hand, exploiting brain-related signals could directly offer rich information about human driver intent,ability and thus allow us to optimize the allocation of authority and reduce the conflict between two agents. A welldesigned human-vehicle interaction interface integrated with human driver psychological and psychobiological characteristics [218], [219] and active capabilities [220] could benefit shared-control tasks by offering much cognition information. Moreover, the feedback from the interaction interface could also show if a poor-designed interface, adverse effects on human driver [221], and control performance [222].Similar to the human-computer interaction design (see pages 29–48 in [223]) from an empirical research perspective,human factors regarding perception sensors (i.e., vision,vestibular and kinesthetic, somatosensory, and auditory, as discussed in Section IV-A-1)), action responders (i.e., limbs and legs, as discussed in Section IV-A-3)), and brain factors(i.e., perception, cognition, and memory, as discussed in Section IV-A-2)) could be carefully considered and implemented based on the methods of human-robot shared control (see the review literature in [224], [225]). One of the typical applications is to use brain-related signals to enhance shared-control performance through the brain-vehicle interface [226]–[229]. Various brain-vehicle interfaces with different capabilities (e.g., adaptive brain-vehicle interface[230]) have been designed based on EEG signals [231]). More literature refers to the review literature [21], [232].

    Moreover, human factors (e.g., the way to shared information with the human driver) can influence the transitions of A2RC. An inefficient design of shared control interface would cause conflicts between human drivers and automated driving agent, even the catastrophic consequences.Eriksson et al. [233] investigated different ways to support driver decision-making during automation-to-manual transitions in a take-over scenario. Most related works have been comprehensively discussed in the recent review paper [234].However, it is still not entirely clear about the influence of different ways (such as torque, steering angle, and vibration)to information transition between a human driver and automated driving agents on the performance during shared control.

    VI. Discussion and Conclusion

    The motivation for this review paper is to illustrate how to model a driver-vehicle shared control system and understand the challenges and opportunities for highly automated vehicles with human drivers still retained in the control loop. We have discussed the architectures of driver-vehicle shared control systems, the approaches to modeling the complex systems,and the future challenges and opportunities. We have provided a survey of the progress over the past decades in drivervehicle shared control technologies. In order to understand the complex driver-vehicle systems, we decoupled it into different subsystems and summarized how to model them by reviewing the state-of-the-art literature. Finally, we have provided discussions on the challenges and opportunities in this field.While advanced driver assistance systems have been developed and introduced over the past decade, a deeper and more holistic understanding of the relationship between human drivers and automated driving agents and the way that human drivers cognitively interact with the driving environment will remain an active area of research in next few years.

    国产真实乱freesex| 久久久久久久久久成人| 亚洲激情五月婷婷啪啪| 麻豆国产97在线/欧美| 一区二区三区乱码不卡18| 国产精品国产三级国产专区5o | 亚洲精品aⅴ在线观看| 91av网一区二区| 激情 狠狠 欧美| 好男人在线观看高清免费视频| 国产在线一区二区三区精 | 亚洲欧洲日产国产| 亚洲人与动物交配视频| 日韩在线高清观看一区二区三区| av在线观看视频网站免费| 国产亚洲午夜精品一区二区久久 | 一区二区三区四区激情视频| 最近手机中文字幕大全| 观看免费一级毛片| 少妇高潮的动态图| 亚洲av成人精品一二三区| 日韩成人av中文字幕在线观看| 人妻系列 视频| 丝袜喷水一区| 91久久精品电影网| 日韩欧美三级三区| 国产精品嫩草影院av在线观看| 亚洲av电影不卡..在线观看| 久久鲁丝午夜福利片| 黄片无遮挡物在线观看| 两个人视频免费观看高清| 女人被狂操c到高潮| 97超碰精品成人国产| 国产乱人视频| 偷拍熟女少妇极品色| 免费黄色在线免费观看| 久久久久久久久久成人| 亚洲欧美成人综合另类久久久 | 最近中文字幕2019免费版| 日韩精品有码人妻一区| 亚洲国产最新在线播放| 日韩欧美国产在线观看| 午夜福利在线观看免费完整高清在| 国产伦精品一区二区三区视频9| 久久精品久久精品一区二区三区| 亚洲美女搞黄在线观看| 九九在线视频观看精品| 长腿黑丝高跟| 久久久久性生活片| 天天一区二区日本电影三级| 亚洲伊人久久精品综合 | kizo精华| 自拍偷自拍亚洲精品老妇| 国产亚洲最大av| 51国产日韩欧美| 精品国内亚洲2022精品成人| 国产女主播在线喷水免费视频网站 | 97人妻精品一区二区三区麻豆| 我要看日韩黄色一级片| 晚上一个人看的免费电影| 伦精品一区二区三区| av播播在线观看一区| 日韩强制内射视频| 成人二区视频| .国产精品久久| 精品一区二区免费观看| 1024手机看黄色片| 亚洲国产日韩欧美精品在线观看| 少妇丰满av| 女人十人毛片免费观看3o分钟| 欧美另类亚洲清纯唯美| 亚洲国产高清在线一区二区三| 免费电影在线观看免费观看| 精品人妻熟女av久视频| 搞女人的毛片| 国产真实伦视频高清在线观看| 日本免费a在线| 成人鲁丝片一二三区免费| av在线亚洲专区| 国产精品综合久久久久久久免费| 亚洲精品aⅴ在线观看| 免费在线观看成人毛片| 麻豆成人午夜福利视频| 一级黄片播放器| 国产麻豆成人av免费视频| 亚洲精品456在线播放app| 又黄又爽又刺激的免费视频.| 熟妇人妻久久中文字幕3abv| 国产极品精品免费视频能看的| 国产伦精品一区二区三区视频9| 亚洲aⅴ乱码一区二区在线播放| 中文字幕精品亚洲无线码一区| 乱人视频在线观看| 久久6这里有精品| 亚洲国产精品成人久久小说| 国产91av在线免费观看| 亚洲成av人片在线播放无| 亚洲电影在线观看av| 亚洲国产精品专区欧美| 亚洲成人久久爱视频| 国产精品人妻久久久久久| 国产精品国产三级专区第一集| 深爱激情五月婷婷| 欧美成人精品欧美一级黄| 精品酒店卫生间| 亚洲伊人久久精品综合 | 最近最新中文字幕大全电影3| 亚洲,欧美,日韩| 九草在线视频观看| 国内精品一区二区在线观看| 国产免费一级a男人的天堂| 久久99热这里只有精品18| 青青草视频在线视频观看| 国产老妇伦熟女老妇高清| 亚洲国产精品久久男人天堂| 国产精品人妻久久久久久| 永久网站在线| 日韩精品有码人妻一区| 国产精品精品国产色婷婷| 欧美色视频一区免费| 在线播放国产精品三级| 欧美日韩国产亚洲二区| 国产免费一级a男人的天堂| 99久久中文字幕三级久久日本| 超碰av人人做人人爽久久| 亚洲人与动物交配视频| 91在线精品国自产拍蜜月| 欧美日韩精品成人综合77777| 亚洲成人中文字幕在线播放| 丰满乱子伦码专区| 美女cb高潮喷水在线观看| 亚洲av不卡在线观看| 中文字幕av在线有码专区| 成人毛片a级毛片在线播放| 久久99热这里只有精品18| 日本av手机在线免费观看| 精品无人区乱码1区二区| 亚洲人成网站在线播| av在线亚洲专区| 永久免费av网站大全| 最近最新中文字幕免费大全7| 色哟哟·www| 18禁裸乳无遮挡免费网站照片| 国产麻豆成人av免费视频| 国产在线一区二区三区精 | 熟妇人妻久久中文字幕3abv| 日韩在线高清观看一区二区三区| 免费观看在线日韩| 三级男女做爰猛烈吃奶摸视频| 91av网一区二区| 久久久精品94久久精品| 亚洲欧美清纯卡通| 午夜激情福利司机影院| 亚洲人成网站在线播| 国产伦一二天堂av在线观看| 寂寞人妻少妇视频99o| 国产精品久久久久久精品电影| 乱人视频在线观看| 简卡轻食公司| 国产麻豆成人av免费视频| 亚洲精品乱久久久久久| 我要看日韩黄色一级片| 国产人妻一区二区三区在| 春色校园在线视频观看| 国产91av在线免费观看| 久久久久久久久久成人| 午夜日本视频在线| 国产成人精品婷婷| 一本久久精品| 精品人妻一区二区三区麻豆| 黄色配什么色好看| 国产一区二区三区av在线| 男女边吃奶边做爰视频| 国产精品嫩草影院av在线观看| 欧美3d第一页| 一级毛片aaaaaa免费看小| 亚洲欧美日韩东京热| 日本免费一区二区三区高清不卡| 高清日韩中文字幕在线| 99国产精品一区二区蜜桃av| 欧美xxxx黑人xx丫x性爽| a级毛色黄片| 一级爰片在线观看| 久久精品国产亚洲av涩爱| 男的添女的下面高潮视频| 久久精品夜夜夜夜夜久久蜜豆| 日本猛色少妇xxxxx猛交久久| 99热全是精品| av.在线天堂| 91久久精品国产一区二区三区| 国产v大片淫在线免费观看| 在线天堂最新版资源| 久久久精品欧美日韩精品| 中文字幕亚洲精品专区| 国产精品乱码一区二三区的特点| 久久欧美精品欧美久久欧美| 少妇丰满av| 一夜夜www| 成人av在线播放网站| 成人欧美大片| 亚洲精品自拍成人| 国产高清视频在线观看网站| av在线播放精品| 国产成人aa在线观看| 中文在线观看免费www的网站| 国产探花极品一区二区| 男女啪啪激烈高潮av片| 欧美精品国产亚洲| 久久99热这里只频精品6学生 | 中国国产av一级| 亚洲最大成人手机在线| 干丝袜人妻中文字幕| 午夜免费男女啪啪视频观看| 三级男女做爰猛烈吃奶摸视频| 国产精品无大码| 亚洲精品国产成人久久av| 国产片特级美女逼逼视频| 亚洲人成网站高清观看| 熟妇人妻久久中文字幕3abv| 永久网站在线| 欧美日韩精品成人综合77777| 亚洲av免费在线观看| 男插女下体视频免费在线播放| 欧美成人精品欧美一级黄| 女人久久www免费人成看片 | 亚洲欧美中文字幕日韩二区| 精品人妻一区二区三区麻豆| 三级男女做爰猛烈吃奶摸视频| 九九久久精品国产亚洲av麻豆| 狂野欧美激情性xxxx在线观看| 成人av在线播放网站| 一级毛片aaaaaa免费看小| 午夜福利在线观看免费完整高清在| www日本黄色视频网| 我要搜黄色片| 亚洲人与动物交配视频| 欧美一级a爱片免费观看看| 一个人看的www免费观看视频| 亚洲欧美日韩无卡精品| 久久久精品94久久精品| 在现免费观看毛片| 又爽又黄a免费视频| 国产熟女欧美一区二区| 亚洲美女视频黄频| 一级毛片久久久久久久久女| 最近的中文字幕免费完整| 国产 一区 欧美 日韩| av.在线天堂| 日日撸夜夜添| 性插视频无遮挡在线免费观看| 欧美另类亚洲清纯唯美| av专区在线播放| 只有这里有精品99| 99久久无色码亚洲精品果冻| 国产精品蜜桃在线观看| 高清视频免费观看一区二区 | 深夜a级毛片| 国产免费视频播放在线视频 | 日韩亚洲欧美综合| 国产乱人偷精品视频| 欧美激情久久久久久爽电影| 亚洲综合色惰| 午夜免费激情av| 韩国高清视频一区二区三区| 99热精品在线国产| av天堂中文字幕网| 久久这里只有精品中国| 免费电影在线观看免费观看| 国产黄片视频在线免费观看| 中文字幕免费在线视频6| 国产伦一二天堂av在线观看| 汤姆久久久久久久影院中文字幕 | 人人妻人人澡欧美一区二区| 岛国在线免费视频观看| 亚洲av电影不卡..在线观看| 一个人看的www免费观看视频| 婷婷六月久久综合丁香| 性插视频无遮挡在线免费观看| 女人被狂操c到高潮| 夫妻性生交免费视频一级片| 99久久精品热视频| 一夜夜www| 麻豆乱淫一区二区| 免费无遮挡裸体视频| АⅤ资源中文在线天堂| 免费看av在线观看网站| 午夜免费男女啪啪视频观看| 成人三级黄色视频| 一个人观看的视频www高清免费观看| 天堂影院成人在线观看| av福利片在线观看| 国产一区二区在线av高清观看| 日韩亚洲欧美综合| 久久久欧美国产精品| www日本黄色视频网| 性色avwww在线观看| 国产在线男女| ponron亚洲| 精品午夜福利在线看| 一级毛片我不卡| 亚洲国产精品久久男人天堂| 免费观看在线日韩| 中国国产av一级| 老司机福利观看| 免费无遮挡裸体视频| 精品一区二区免费观看| 午夜免费激情av| 亚洲三级黄色毛片| 免费看a级黄色片| 日本猛色少妇xxxxx猛交久久| 超碰97精品在线观看| 全区人妻精品视频| 国产高清国产精品国产三级 | 国产成人精品一,二区| 寂寞人妻少妇视频99o| 91久久精品国产一区二区三区| 成人二区视频| 99久久九九国产精品国产免费| 亚洲av福利一区| 国产精品一二三区在线看| 尾随美女入室| 成人三级黄色视频| 丰满少妇做爰视频| 午夜精品国产一区二区电影 | 亚洲18禁久久av| 国产成人aa在线观看| a级一级毛片免费在线观看| 亚洲av日韩在线播放| 亚洲精品色激情综合| 久久人人爽人人片av| 看十八女毛片水多多多| 久久精品影院6| 国产黄片视频在线免费观看| 午夜精品国产一区二区电影 | 淫秽高清视频在线观看| 麻豆av噜噜一区二区三区| 欧美潮喷喷水| 国产免费一级a男人的天堂| 日日干狠狠操夜夜爽| 亚洲丝袜综合中文字幕| 国产av在哪里看| 国产高清有码在线观看视频| 国产精品国产三级国产专区5o | 人妻系列 视频| 91午夜精品亚洲一区二区三区| 精品欧美国产一区二区三| av.在线天堂| 一本久久精品| 人体艺术视频欧美日本| 午夜福利视频1000在线观看| 久久精品91蜜桃| 性色avwww在线观看| 三级男女做爰猛烈吃奶摸视频| 欧美成人a在线观看| 国产毛片a区久久久久| 一区二区三区四区激情视频| 国产精品.久久久| 日日摸夜夜添夜夜爱| 亚洲国产精品合色在线| 免费搜索国产男女视频| 亚洲成人av在线免费| 欧美xxxx黑人xx丫x性爽| 国产精品1区2区在线观看.| 毛片一级片免费看久久久久| 人体艺术视频欧美日本| 午夜精品在线福利| 中文天堂在线官网| 99热全是精品| 国产 一区精品| 99久国产av精品| 五月伊人婷婷丁香| 99热这里只有精品一区| 久久久久久久久大av| 尤物成人国产欧美一区二区三区| 国产精品,欧美在线| 天堂影院成人在线观看| 亚洲自偷自拍三级| 久久久久久久亚洲中文字幕| 亚洲精品成人久久久久久| 国产精华一区二区三区| 成人欧美大片| 亚洲自拍偷在线| 18禁在线无遮挡免费观看视频| 乱码一卡2卡4卡精品| 男的添女的下面高潮视频| av视频在线观看入口| 国产精品人妻久久久久久| 色综合站精品国产| 尾随美女入室| 欧美zozozo另类| 国产精品.久久久| 好男人视频免费观看在线| 精品不卡国产一区二区三区| 男人和女人高潮做爰伦理| 国产成人精品久久久久久| 在现免费观看毛片| 亚洲激情五月婷婷啪啪| 久久久国产成人精品二区| 少妇人妻一区二区三区视频| 高清毛片免费看| 99热网站在线观看| 精品国内亚洲2022精品成人| 亚洲最大成人av| 欧美日韩国产亚洲二区| 日韩av在线大香蕉| 免费av不卡在线播放| 一本久久精品| 国内少妇人妻偷人精品xxx网站| 在线免费观看不下载黄p国产| 亚洲av中文av极速乱| 国产视频内射| 国产一区二区在线观看日韩| 中文字幕人妻熟人妻熟丝袜美| 国产精品国产三级国产av玫瑰| 成人综合一区亚洲| 青春草亚洲视频在线观看| 国产午夜精品论理片| 嫩草影院新地址| 伦精品一区二区三区| 少妇的逼好多水| 国产成人a区在线观看| 有码 亚洲区| 真实男女啪啪啪动态图| 九九久久精品国产亚洲av麻豆| 99久国产av精品| 免费看a级黄色片| 成人性生交大片免费视频hd| a级一级毛片免费在线观看| 久久这里只有精品中国| 久久人人爽人人片av| 亚洲精品aⅴ在线观看| 久久99热这里只频精品6学生 | 国内精品美女久久久久久| 91精品伊人久久大香线蕉| 中文资源天堂在线| 床上黄色一级片| 日本一本二区三区精品| 菩萨蛮人人尽说江南好唐韦庄 | 久久精品夜夜夜夜夜久久蜜豆| 国产在视频线在精品| 国产男人的电影天堂91| 精品久久久久久久人妻蜜臀av| 我要搜黄色片| 成人亚洲欧美一区二区av| 精品欧美国产一区二区三| 午夜日本视频在线| 欧美极品一区二区三区四区| 99久久中文字幕三级久久日本| 精品国内亚洲2022精品成人| 青青草视频在线视频观看| 亚洲色图av天堂| 国产精华一区二区三区| 午夜福利在线在线| 成人亚洲精品av一区二区| av免费在线看不卡| 欧美变态另类bdsm刘玥| 欧美zozozo另类| 亚洲中文字幕日韩| 少妇熟女欧美另类| 亚洲av成人精品一二三区| 国产成人精品一,二区| 国产视频首页在线观看| 性插视频无遮挡在线免费观看| 夜夜爽夜夜爽视频| 国产私拍福利视频在线观看| 国产精品久久电影中文字幕| 亚洲国产精品sss在线观看| 99热全是精品| 亚洲精品乱码久久久久久按摩| 好男人在线观看高清免费视频| 国产成人精品久久久久久| 在线免费观看不下载黄p国产| 日韩欧美精品v在线| 日本三级黄在线观看| 国产高清视频在线观看网站| 国产精品不卡视频一区二区| 国产精品久久视频播放| 欧美精品一区二区大全| 婷婷色av中文字幕| 高清日韩中文字幕在线| 国产成人午夜福利电影在线观看| 亚洲aⅴ乱码一区二区在线播放| 免费观看在线日韩| 国产精品一二三区在线看| 草草在线视频免费看| 啦啦啦观看免费观看视频高清| 男女国产视频网站| 成年版毛片免费区| 能在线免费观看的黄片| 91av网一区二区| videos熟女内射| 国产一区二区亚洲精品在线观看| 欧美激情久久久久久爽电影| 国产老妇伦熟女老妇高清| av播播在线观看一区| av线在线观看网站| 欧美一区二区国产精品久久精品| 色综合色国产| 久久国内精品自在自线图片| 蜜臀久久99精品久久宅男| 欧美高清性xxxxhd video| 亚洲熟妇中文字幕五十中出| 国产伦在线观看视频一区| av国产久精品久网站免费入址| 在线观看av片永久免费下载| 亚洲人成网站在线播| 丰满人妻一区二区三区视频av| 国产高清三级在线| 长腿黑丝高跟| 久久久精品大字幕| 我的女老师完整版在线观看| 亚洲美女搞黄在线观看| 午夜a级毛片| 欧美成人午夜免费资源| 亚洲欧洲国产日韩| 国产一级毛片在线| 久久国产乱子免费精品| 成人鲁丝片一二三区免费| 深夜a级毛片| 极品教师在线视频| 最新中文字幕久久久久| 精品久久久久久成人av| 草草在线视频免费看| 国产av一区在线观看免费| 99久久中文字幕三级久久日本| 国产午夜精品一二区理论片| 91精品伊人久久大香线蕉| 午夜福利高清视频| 中文字幕免费在线视频6| 有码 亚洲区| 成年版毛片免费区| 三级毛片av免费| 久久婷婷人人爽人人干人人爱| 国产激情偷乱视频一区二区| 99国产精品一区二区蜜桃av| 国内少妇人妻偷人精品xxx网站| 久久久久久久国产电影| 久久久久久国产a免费观看| 国产黄片美女视频| 亚洲欧美日韩高清专用| 看非洲黑人一级黄片| 欧美97在线视频| 狠狠狠狠99中文字幕| 麻豆久久精品国产亚洲av| 女的被弄到高潮叫床怎么办| 麻豆国产97在线/欧美| 看十八女毛片水多多多| 国产中年淑女户外野战色| 婷婷六月久久综合丁香| kizo精华| 国产精华一区二区三区| 亚洲成人中文字幕在线播放| 免费人成在线观看视频色| 免费搜索国产男女视频| 亚洲av中文av极速乱| 乱码一卡2卡4卡精品| 亚洲av一区综合| 国产一级毛片七仙女欲春2| 晚上一个人看的免费电影| 一边亲一边摸免费视频| 亚洲av二区三区四区| 麻豆乱淫一区二区| 亚洲av免费在线观看| 国产精品乱码一区二三区的特点| 国产人妻一区二区三区在| 亚洲美女视频黄频| 啦啦啦观看免费观看视频高清| 91久久精品国产一区二区三区| 欧美激情国产日韩精品一区| 久久精品久久精品一区二区三区| 不卡视频在线观看欧美| 欧美zozozo另类| 夜夜看夜夜爽夜夜摸| 水蜜桃什么品种好| 中文欧美无线码| 建设人人有责人人尽责人人享有的 | 级片在线观看| 天堂√8在线中文| 免费av观看视频| 天美传媒精品一区二区| 亚洲成色77777| 久久久色成人| 精品国产三级普通话版| 18+在线观看网站| 午夜福利视频1000在线观看| 婷婷色综合大香蕉| 天堂av国产一区二区熟女人妻| 六月丁香七月| 日本与韩国留学比较| 亚洲av成人精品一区久久| 91久久精品国产一区二区成人| 婷婷色麻豆天堂久久 | 一级二级三级毛片免费看| 身体一侧抽搐| 内地一区二区视频在线| 欧美性感艳星| 一区二区三区高清视频在线| 亚洲图色成人| 免费看美女性在线毛片视频| 中文天堂在线官网| 午夜久久久久精精品| 91在线精品国自产拍蜜月| 日韩av在线免费看完整版不卡| 真实男女啪啪啪动态图| 日日摸夜夜添夜夜爱| 日产精品乱码卡一卡2卡三| 日韩视频在线欧美| 嫩草影院入口| 91aial.com中文字幕在线观看| 国产淫语在线视频| 日韩欧美精品免费久久| 欧美日韩精品成人综合77777| 欧美高清成人免费视频www| 老司机影院成人| 菩萨蛮人人尽说江南好唐韦庄 |