• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Topology-preserving flocking of nonlinear agents using optimistic planning

    2015-12-05 07:42:21LucianBUSONIUIrinelConstantinMORARESCU
    Control Theory and Technology 2015年1期

    Lucian BU?SONIU ,Irinel-Constantin MOR?ARESCU

    1.Department of Automation,Technical University of Cluj-Napoca,Memorandumului 28,400114 Cluj-Napoca,Romania;

    2.Universit′e de Lorraine,CRAN,UMR 7039 and CNRS,CRAN,UMR 7039,2 Avenue de la For?et de Haye,Vandoeuvre-l`es-Nancy,France

    Received 22 July 2014;revised 20 January 2015;accepted 20 January 2015

    Topology-preserving flocking of nonlinear agents using optimistic planning

    Lucian BU?SONIU1?,Irinel-Constantin MOR?ARESCU2

    1.Department of Automation,Technical University of Cluj-Napoca,Memorandumului 28,400114 Cluj-Napoca,Romania;

    2.Universit′e de Lorraine,CRAN,UMR 7039 and CNRS,CRAN,UMR 7039,2 Avenue de la For?et de Haye,Vandoeuvre-l`es-Nancy,France

    Received 22 July 2014;revised 20 January 2015;accepted 20 January 2015

    We consider the generalized flocking problem in multiagent systems,where the agents must drive a subset of their state variables to common values,while communication is constrained by a proximity relationship in terms of another subset of variables.We build a flocking method for general nonlinear agent dynamics,by using at each agent a near-optimal control technique from artificial intelligence called optimistic planning.By defining the rewards to be optimized in a well-chosen way,the preservation of the interconnection topology is guaranteed,under a controllability assumption.We also give a practical variant of the algorithm that does not require to know the details of this assumption,and show that it works well in experiments on nonlinear agents.

    Multiagent systems,flocking,optimistic planning,topology preservation

    DOI 10.1007/s11768-015-4107-5

    1 Introduction

    Multi-agent systems such as robotic teams,energy and telecommunication networks,collaborative decision support systems,data mining,etc.appear in many areas of technology.Their component agents usually only have a local,limited view,which means decentralized approaches are necessary to control the overall system.In this decentralized setting,often consensus between the agents is desired,meaning that they must reach agreement on certain controlled variables of interest[1–3].Inspired by the behavior of flocks of birds,researchers studied theflockingvariant of consensus,which only requires consensus on velocities while also using position measurements[3,4].Flocking is highly relevant in e.g.,mobile robot teams[5].

    In this paper,we consider a generalized version of the flocking problem,in which agreement is sought for a subset of agent variables,while other variables define the interconnection topology between the agents.These two subsets may,but need not represent velocities and positions.The communication connections between agents are based on a proximity relationship,in which a connection is active when the agents are closer than some threshold in terms of the connectivity variables.Each agent finds control actions(inputs)using the optimistic planning(OP)algorithm from artificial intelligence[6].OP works with discrete actions,like the consensus method of[7],and finds sequences of actions that are near-optimal with respect to general nonquadratic reward functions,for general nonlinear dynamics.The first major advantage of our technique is this inherited generality:it works for any type of nonlinear agents.A controllability property is imposed that,for any connected state,roughly requires the existence of an input sequence which preserves connectivity.We define agent reward functions with separate agreement and connectivity components,and our main analytical result shows that if the connectivity rewards are sufficiently large,the algorithm will preserve the interconnection topology.In interesting cases,the computational complexity of the flocking problem is not larger than if the agent would solve the agreement-only problem.The theoretical algorithm is restrictive in requiring to know the length of action sequences satisfying the controllability property.We therefore also provide a practical algorithm variant which does not use this knowledge,and validate it in simulation to nonholonomic agents and robot arms[8].In the second problem we illustrate that despite our focus on flocking,the method also works in the full-state consensus case.

    The main novelty of the OP approach compared to existing methods is that it is agnostic to the specific agent dynamics,and so it works uniformly for general nonlinear agents.In particular,our analysis shows that when a solution that preserves the topology exists(in a sense that will be formalized later),then irrespective of the details of the dynamics the algorithm will indeed maintain the topology.Existing topology preservation results are focused on specific types of agents,mainly linear[9,10],[11,Ch.4],or sometimes nonlinear as in e.g.,[12]where the weaker requirement of connectivity is considered.Our practical flocking algorithm exhibits the same generality,whereas existing methods exploit the structure of the specific dynamics targeted to derive predefined control laws,e.g.,for linear double integrators[3],agents with nonlinear acceleration dynamics[13,14],or nonholonomic robots[12,15].The technical contribution allowing us to achieve these results is the exploitation of the OP algorithm,and of its strong near-optimality guarantees.

    The approach presented here is a significantextension of our earlier work[16]:it introduces a new algorithm that is theoretically shown to preserve the topology,and also includes new empirical results for nonholonomic agents.Also related isouroptimistic-optimizationbased approach of[17],which only handles consensus on a fixed graph rather than flocking,and directly optimizes over fixed-length action sequences rather than using planning to exploit the dynamical structure of the control problem.

    The remainder of this paper is organized as follows.After formalizing the problem in Section 2 and explaining OP in Section 3,the two variants of the consensus algorithm and the analysis of the theoretical variant are given in Section 4.Section 5 presents the experimental results and Section 6 concludes the paper.

    List of symbols and notations

    |·| cardinality of argument set;

    nnumber of agents;

    G,V,E,N graph,vertices,edges,neighbors;

    i,jagent indices;

    x,u,fstate,action,dynamics;

    xa,xcagreement states,connectivity states;

    Pcommunication range;

    kabsolute time step;

    Klength of sequence ensuring connectivity;

    udaction sequence of lengthd;

    ρ,vreward function,value(return);

    γ discount factor;

    Δ,Γ agreement reward,connectivity reward;

    β weight of connectivity reward;

    Toptimistic planning budget;

    T,T?,L tree,near-optimal tree,leaves;

    ddepth in the tree(relative time step);

    b,ν upper and lower bound on return;

    κ branching factor of near-optimal tree.

    2 Problem statement

    Consider a set ofnagents with decoupled nonlinear dynamicsxi,k+1=fi(xi,k,ui,k),i=1,...,n,wherexianduidenote the state and action(input)of theith agent,respectively.The agents can be heterogeneous:they can have different dynamics and state or input dimensionality.An agent only has a local view:it can receive information only from its neighbors on an interconnection graph Gk=(V,Ek),which can be time-varying.The set of nodes V={1,...,n}represents the agents,and the edges Ek?V×V are the communication links.Denote by Ni,k={j|(i,j)∈Ek}the set of neighbors of nodeiat stepk.A path through the graph is a sequence of nodesi1,...,iNso that(il,il+1)∈Ek,1llt;N.The graph is connected if there is a path between any pair of nodesi,j.

    The objective can be formalized as

    wherexaselects only those state variables for whichagreementis desired,and·denotes an arbitrary norm.We require of course that the selection produces a vector with the same dimensionality for all agents.When all agents have the same state dimension,xa=x,and Ek=E(a fixed communication graph)we obtain the standard full-state consensus problem[1,2].While our technique can be applied to this case,as will be illustrated in the experiments,in the analytical development we will focus on the flocking problem,where the communication network varies with theconnectivitystate variablesxc.Usually,xaandxcdo notoverlap,being e.g.,the agent’s velocity and position[3],so that velocities mustbe synchronized undercommunication constraints dependent on the position.Specifically,we consider the case where a link is active when the connectivity states of two agents are close:

    For example whenxcis a position this corresponds to the agents being physically closer than some transmission rangeP.

    Our approach requires discretized agent actions.

    Assumption1Agentactions are discretized:ui∈Uiwith|Ui|=Mi.

    Remark 1Certain systems have inherently discrete and finitely-many actions,because they are controlled by switches.When the actions are originally continuous,discretization reduces performance,but the loss is often manageable.Other authors showed interest in multiagent coordination with discretized actions,e.g.,[7].

    Remark 2This is a feasibility assumption:it is difficult to preserve the topology without requiring such a condition.The condition simply means that for any joint state of the system in which an agent is connected to some neighbors,this agent has an action sequence by which it is again connected afterKsteps,if its neighbors do not move.Therefore,if the assumption does not hold and the problem is such that the neighbors do stay still,the agent will indeed lose some connections and topology cannot be preserved.Of course,in general the neighbors will move,but as we will show Assumption 2 is nevertheless sufficient to ensure connectivity.

    K-step controllability properties are thoroughly studied in the literature,e.g.,[18]provide Lie-algebraic conditions to guarantee them.We make a similar assumption in our previous paper[17],where it is however much stronger,requiring thatthe controlis able to move the agent betweenanytwo arbitrary states in a bounded region.With a sufficiently fine action discretization,such an assumption would locally imply Assumption 2 in the present paper.

    When making the assumption,we could also use the following definition for the links:

    so that the agents never gain new neighbors,and only need to stay connected to their initial neighbors.The analysis will also hold in this case,which is important because with(1),askgrows many or all the agents may become interconnected.For simplicity,we use(1)in the sequel.

    3 Background:Optimistic planning for deterministic systems

    Consider a(single-agent)optimal control problem for a deterministic,discrete-time nonlinear systemxd+1=f(xd,ud)with statesxand actionsu.Define an infinitelylong action sequenceu∞=(u0,u1,...)and its truncation todinitial actions,u d=(u0,...,ud?1).Given an initial statex0,the return of a sequence is

    Optimistic planning for deterministic systems(OP)[6,21]explores a tree representation of the possible action sequences from the current system state,as illustrated in Fig.1.It requires a discrete action spaceU=?u1,...,uM?;recall Assumption 1,which ensures this is true forouragents.OP startswith a rootnode representing the empty sequence,and iteratively expandsTwell-chosen nodes.Expanding a node addsMnew children nodes for all possible discrete actions.Each node at some depthdis reached via a unique path through the tree,associated to a unique action sequenceu dof lengthd.We will denote the nodes by their corresponding action sequences.Denote also the current tree by T,and its leaves by L(T).

    For a leaf nodeu d,the following gives an upper bound on the returns of all infinite action sequences having in common the initial subsequence up tou d:

    where ν(u d)is a lower bound.These properties hold because all the rewards at depths larger thandare in[0,1].

    Fig.1 Illustration of an OP tree T.Nodes are labeled by actions,arcs represent transitions and are labeled by the resulting states and rewards.Subscripts are depths,superscripts index the M possible actions/transitions from a node(here,M=2).The leaves are enclosed in a dashed line,while the gray thick path highlights a sequence.

    Algorithm 1(Optimistic planning for deterministic systems)Initialize tree:T←?empty sequence u0?;for t=1,...,T do find optimistic leaf:u?← arg max b(u),add to T the children of u?,labeled by u1,...,uM end for;return u?u∈L(T)d?,where u?d?=arg maxν(u).u∈L(T)

    Usually OP and its analysis are developed for timeinvariant reward functions[6,21],such as the quadratic reward exemplified above.However,thisfactis notused in the development,which therefore entirely carriesover to the time-varying case explained here.We provide the algorithm and results directly in the time-varying case,since this will be useful in the consensus context.

    To characterize the complexity of finding the optimal sequence from a given statex,we use the asymptotic branching factor of the near-optimal subtree:

    Asequenceudis said to beε-optimalwhenv??v(ud)ε.The upcoming theorem is a consequence of the analysis in[6,21].It is given here in a form that brings out the role of the sequence length,useful later.Part i)of the theorem shows that OP returns a long,near-optimal sequence,while part ii)quantifies this length and nearoptimality,via branching factor κ.

    Theorem 1When OP is called with budgetT:

    i)The lengthd?of the sequenceu?

    The smaller κ,the better OP does.The best case is κ=1,obtained e.g.,when a single sequence always obtains rewards of 1,and all the other rewards on the tree are 0.In this case the algorithm must only develop this sequence,and suboptimality decreases exponentially.In the worst case,κ=M,obtained e.g.,when all the sequences have the same value,and the algorithm must explore the complete tree in a uniform fashion,expanding nodes in order of their depth.

    4 Flocking algorithm and analysis

    The OP-based approach to the flocking problem in Section 2 works as follows.At every time stepk,a local optimal control problem is defined for each agenti,using information locally available to it.The goal in this problem is to align the agreement statesxawith those of the neighbors Ni,k,while maintaining the connection topology by staying close to them in terms ofxc.OP is used to near-optimally solve this control problem,and an initial subsequence of the sequence returned is applied by the agent.Then the system evolves,and the procedure is applied again,for the new states and possibly changed graph.

    To construct its optimal control problem,each agent needs the predicted behavior of its neighbors.Here,agents will exchange the predicted state sequences resulting from the near-optimalaction sequences returned by OP.Because the agents must act at the same time,how they exchange predictions is nontrivial.If predictions do not match,acoordinationproblem may arise where mismatching actions are applied.Coordination is a difficult challenge in multi-agent systems and is typically solved in model-predictive control by explicit,iterative negotiation over successive local solutions,e.g.,[22].However,it is unlikely that the agents can computationally afford to repeatedly communicate and reoptimize their solutions at every step.Thus we adopt a sequential communication procedure in which agents optimize once per step,similar to the procedure for distributed MPC in[23].We show in Section 4.1 that connectivity can be guaranteed despite this one-shot solution.

    To implement the sequential procedure,each agent needs to know its indexias well as the indices of its neighbors.One way to ensure this is an initial,centralized assignment of indices to the agents.Agentiwaits until the neighborsjwithjlt;ihave solved their local optimalcontrolproblemsand found theirpredicted state sequences.These agents communicate their predictions toi.Forjgt;i,agenticonstructs other predictions as described later.Agentioptimizes its own behavior while coordinating with the predictions.It then sends its own,newly computed prediction to neighborsjgt;i.

    The local optimal control problem of agentiis then defined using the reward function:

    In the implementation,if the agents have their neighbors’models,they could also exchange predicted action sequences instead of states.Since actions are discrete and states usually continuous,this saves some bandwidth at the cost of extra computation to resimulate the neighbor’s transitions up to the prediction length.In any case,itshould be noted thatagents donotoptimize over the actions of their neighbors,so complexity does not directly scale with the number of neighbors.

    So far,we have deliberately left open the specific form of the rewards and predictions for neighborsjgt;i.Next,we instantiate them in a theoretical algorithm for which we guarantee the preservation of the interconnection topology and certain computational properties.However,this theoretical variant has shortcomings,so we additionally present a different instantiation which is more suitable in practice and which we later show works well in experiments.

    4.1 A theoretical algorithm with guaranteed topology preservation

    Our aim in this section is to exploit Assumption 2 to derive an algorithm that preserves the communication connections.We first develop the flocking protocol for each agent,shown as Algorithm 2.Our analysis proceeds by showing that,if sequences preserving the connections exist at a given step,the rewards can be designed to ensure that the algorithm will indeed find one such sequence(Lemma 1).This property is then used to prove topology preservation in closed loop,in Theorem 2.Finally,Theorem 3 shows an interesting computational property of the algorithm:under certain conditions the extra connectivity reward does notincrease the complexity from the case where only agreement would be required.

    The agreement reward is left general,but to fix ideas,it could be,for instance,

    Algorithm 2(OP flocking at agent i–theoretical variant)Set initial prediction x?1i to an empty sequence;for?=0,1,2,...do current step is k←?K,exchange state at k with all neighbors j∈N i,k,send xk?1i to jlt;i,wait to receive new predictions?xi,kj from all jlt;i,form predictions?xi,kj for jgt;i,run OP with(5)and(6),obtaining uki and xki,send xki to jgt;i,execute K actions uki,0,...,uki,K?1 in open loop end for.

    Moving on to the analysis now,we first show that when it is possible,each agent preserves connectivity with respect to the predicted states of its neighbors.

    Putting the local guarantees together,we have topology preservation for the entire system,as follows.

    Theorem 2Take β andTas in Lemma 1,then under Assumption 2 and if the graph is initially connected,Algorithm 2 preserves the connections at any stepk=?K.

    ProofThe intuition is very simple:each agentiwill move so as to preserve connectivity with thepreviousstate of any neighborjgt;i,and then in turnjwill move while staying connected with theupdatedstate ofi,which is what is required.However,since Assumption 2 requires connectivity to hold globally for all neighbors,the formal proof is somewhat technical.

    Theorem 2 guarantees that thetopologyis preserved when the initial agent states correspond to a connected network.However,this result does not concern the stability of theagreement.In practice,we solve the agreement problem by choosing appropriately the rewards Δ,such as in(7),so that by maximizing the discounted returnsthe agentsachieve agreement.In Section 5,we illustrate thatthis approach performswellin experiments.Note that Theorem 2 holds whether the graph is defined with(1)or(2).

    We will compare performance in the original problem with that in the agreement problem.

    Theorem 3Assumev?=v?u.For OP applied to theoriginalproblem,the near-optimality bounds of Theorem 1 ii)hold with the branching factor κaof theagreementproblem.

    ProofWe startwith a slightmodification to the analysis of OP.For any problem,define the set:

    Therefore,finally,

    Theorem 3 can be interpreted as follows.If the unconstrained optimal solution would have naturally satisfied connectivity(which is not unreasonable),adding the constraint does not harm the performance of the algorithm,so that flocking is as easy as solving only the agreement problem.This is a nice property to have.

    4.2 A practical algorithm

    Algorithm 2 has an important shortcoming in practice:it requires knowing a value ofKfor which Assumption 2 is satisfied.Further,keeping predictions constant forjgt;iis safe,but conservative,since better predictions are usually available:those made by the neighbors at previous steps,which may not be expected to change much,e.g.,when a steady state is being approached.

    SinceKis unknown,the agent will try preserving connectivity ateverystep,with as many neighbors as possible:

    For the links,definition(1)is used,since old neighbors may be lost but the graph may still remain connected due to new neighbors.Therefore,the aim here is only connectivity,weakerthan topology preservation.Forthe agreement component,(7)is employed.Algorithm 3 summarizes the resulting protocol for generic agenti.

    A l g o r i t h m 3(O P f l o c k i n g a t a g e n t i–p r a c t i c a l v a r i a n t)S e t i n i t i a l p r e d i c t i o n x?1i t o a n e m p t y s e q u e n c e;f o r s t e p k=0,1,2,...d o e x c h a n g e s t a t e a t k w i t h a l l n e i g h b o r s j∈Ni,k,s e n d xk?1i t o jlt;i,r e c e i v e xk?1j f r o m jgt;i,w a i t t o r e c e i v e n e w p r e d i c t i o n s ? xi,kj f r o m a l l jlt;i,f o r m p r e d i c t i o n s ? xi,kj f o r jgt;i,r u n O P w i t h(5)a n d(8),o b t a i n i n g uki a n d xki,s e n d xki t o jgt;i,e x e c u t e a c t i o n uk i,0 e n d f o r.

    The main advantage of our approach,in both Algorithm 2 and Algorithm 3,is the generality of the agent dynamics it can address.This generality comes at the cost of communicating sequences of states,introducing a dependence of the performance on the action discretization,and a relatively computationally involved algorithm.The time complexity of each individual OP application is between O(TlogT)and O(T2)depending onκ.The overallcomplexity forallagents,ifthey run OP in parallel as soon as the necessary neighbor predictions become available,is largerby a factorequalto the length of the longest path from anyito anyjgt;i.Depending on the current graph this length may be significantly smaller than the number of agentsn.

    5 Experimental results

    The proposed method is evaluated in two problems with nonlinear agent dynamics.The first problem concerns flocking fora simple type ofnonholonomic agents,where we also study the influence of the tuning parameters of the method.In the second experiment,full-state consensus for two-link robot arms is sought.This experiment illustrates that the algorithm can on the one hand handle rather complicated agent dynamics,and on the other hand that it also works for standard consensus on a fixed graph,even though our analytical focus was placed on the flocking problem.

    While both types of agents have continuous-time underlying dynamics,they are controlled in discrete time,as is commonly done in practical computer-controlled systems.The discrete-time dynamics are then the result of integrating the continuous-time dynamics with zero-order-hold inputs.Then,in order for the analysis to hold for Algorithm 2,Assumption 2 must be satisfied by these discretized dynamics.Note that in practice we apply Algorithm 3,and the numerical integration technique introduces modelerrorsthatouranalysis does not handle.

    5.1 Flocking of nonholonomic agents

    Consider homogeneous agents that evolve on a plane and have the state vectorx=[X Y vθ]withX,Ythe position on the plane(m),vthe linear velocity(m/s),and θ the orientation(rad).The control inputs are the rate of changeaof the velocity and ω of the orientation.The discrete-time dynamics are

    where Euler discretization with sampling timeTswas employed.The aim is to agree onxa=[vθ]T,which represent the velocity vector of the agent,while maintaining connectivity on the plane by keeping the distances between the connectivity statesxc=[X Y]Tbelow the communication rangeP.

    The specific multiagent system we experiment with consists of 9 agents initially arranged on a grid with diverging initial velocities,see Fig.2(a).Their initial communication graph has some redundant links.In the reward function,β=0.5 so that agreement and connectivity rewards have the same weight,and the agreement reward is(7)with the distance measure being a 2-norm weighted so that it saturates to 1 at a distance 5 between the agreement states.The range isP=5.The sampling time isTs=0.25 s.

    Fig.2 shows that the OP method preserves connectivity while achieving flocking,up to errors due mainly to the discretized actions.The discretized action set was{?0.5,0,0.5}m/s2× {?π/3,0,π/3}rad/s,and the planning budget of each agent isT=300 node expansions.For all the experiments,the discount factor γ is set to 0.95,so that long-term rewards are considered with significant weight.

    Fig.2 Results for nonholonomic agents.(a)Initial configuration,with the agents shown as colored dots,their initial velocities and orientations symbolized by the thick lines,and their initial graph with thin gray lines,(b)trajectories on the plane,also showing the final configuration of the agents,and(c)evolution of agreement variables.

    The following budgets are used:T=25,50,75,100,200,...,600,and the length of the predictions is not limited.As shown in Fig.3 and as expected from the theoreticalguarantees ofOP,disagreementlargely decreases withTalthough the decrease is not monotonic1See[24],footnote 4 for an example showing how nonmonotonicity can happen..

    Fig.3 Influence of the expansion budget.

    The influence of the prediction length is studied for fixedT=300,by takingL=0,1,3,4 and then allowing full predictions2In effect,predictions with this budget do not exceed length 4 so the last two results will be identical..Fig.4 indicates that performance is not monotonic inL,and medium-length predictions are better in this experiment.While it is expected that too long predictions will not increase performance since they will rarely be actually be followed,the good results for communicating just the current state without any prediction are more surprising,and need to be studied further.

    Fig.4 Influence of the maximal prediction length.

    5.2 Consensus of robotic arms

    Three robots are connected on a fixed undirected communication graph in which robot 1 communicates with both 2 and 3,but 2 and 3 are not connected.The initial angular positions are taken random with zero initial velocities,see Fig.5.

    Fig.5 Leaderless consensus of multiple robotic arms:angles and angular velocities for the two links,overimposed for all the robots.Angles wrap around in the interval[?π,π).

    The distance measure is the squared Euclidean distance,weighted so that the angular positions are given priority.The discretized actions are{?1.5,0,1.5}N·m×{?1,0,1}N·m,and the budget of each agent isT=400.Consensus is achieved without problems.

    6 Conclusions

    We have provided a flocking technique based on optimistic planning(OP),which under appropriate conditions is guaranteed to preserve the connectivity topology of the multiagent system.A practical variant of the technique worked well in simulation experiments.

    An important future step is to develop guarantees also on the agreement component of the state variable.This is related to the stability of the near-optimal control produced by OP,and since the objective function is discounted such a stability property is a big open question in the optimal control field[26].It would also be interesting to apply optimistic methods to other multiagent problems such as gossiping or formation control.

    [1]R.Olfati-Saber,J.A.Fax,R.M.Murray.Consensus and cooperation in networked multiagent systems.Proceedings of the IEEE,2007,95(1):215–233.

    [2]W.Ren,R.W.Beard.Distributed Consensus in Multi-vehicle Cooperative Control:Theory and Applications.Communications and Control Engineering.Berlin:Springer,2008.

    [3]R.Olfati-Saber.Flocking for multi-agent dynamic systems:Algorithms and theory.IEEE Transactions on Automatic Control,2006,51(3):401–420.

    [4]H.G.Tanner,A.Jadbabaie,G.J.Pappas.Flocking in fixed and switching networks.IEEE Transactions on Automatic Control,2007,52(5):863–868.

    [5]W.Dong.Flocking of multiple mobile robots based on backstepping.IEEE Transactions on Systems,Man,and Cybernetics–Part B:Cybernetics,2011,41(2):414–424.

    [6]J.-F.Hren,R.Munos.Optimistic planning of deterministic systems.Proceedings 8th European Workshop on Reinforcement Learning,Villeneuve d’Ascq,France:Springer,2008:151 – 164.

    [7]C.De Persis,P.Frasca.Robust self-triggered coordination with ternary controllers.IEEE Transactions on Automatic Control,2013,58(12):3024–3038.

    [8]J.Mei,W.Ren,G.Ma.Distributed coordinated tracking with a dynamic leader for multiple Euler-Lagrange systems.IEEE Transactions on Automatic Control,2011,56(6):1415–1421.

    [9]M.M.Zavlanos,G.J.Pappas.Distributed connectivity control of mobile networks.IEEE Transactions on Robotics,2008,24(6):1416–1428.

    [10]M.Fiacchini,I.C.Mor?arescu.Convex conditionson decentralized control for graph topology preservation.IEEE Transactions on Automatic Control,2014,59(6):1640–1645.

    [11]F.Bullo,J.Cort′es,S.Martinez.Distributed Control of Robotic Networks:A Mathematical Approach to Motion Coordination Algorithms.Princeton:Princeton University Press,2009.

    [12]J.Zhu,J.Lu,X.Yu.Flocking of multi-agent non-holonomic systems with proximity graphs.IEEE Transactions on Circuits and Systems,2013,60(1):199–210.

    [13]H.Su,G.Chen,X.Wang,et al.Adaptive second-order consensus of networked mobile agents with nonlinear dynamics.Automatica,2011,47(2):368–375.

    [14]J.Zhou,X.Wu,W.Yu,et al.Flocking of multi-agent dynamical systems based on pseudo-leader mechanism.Systemsamp;Control Letters,2012,61(1):195–202.

    [15]H.Tanner,A.Jadbabaie,G.Pappas.Flocking in teams of nonholonomic agents.Cooperative Control.V.Kumar,N.Leonard,A.Morse(eds.),Berlin:Springer,2005:458–460.

    [16]L.Bu?soniu,C.Mor?arescu.Optimistic planning for consensus.Proceedings of the American Control Conference,Washington D.C.:IEEE,2013:6735–6740.

    [17]Lucian Bu?soniu,C.Mor?arescu.Consensus for blackbox nonlinear agents using optimistic optimization.Automatica,2014,50(4):1201–1208.

    [18]B.Jakubczyk,E.D.Sontag.Controllability of nonlinear discretetime systems:A Lie-algebraic approach.SIAM Journal of Control and Optimization,1990,28(1):1–33.

    [19]R.S.Sutton,A.G.Barto.ReinforcementLearning:An Introduction.Cambridge,MA:MIT Press,1998.

    [20]F.Lewis,D.Liu,eds.Reinforcement Learning and Approximate Dynamic Programming for Feedback Control.Hoboken:John Wileyamp;Sons,2012.

    [21]R.Munos.The optimistic principle applied to games,optimization and planning:Towards foundations of Monte-Carlo tree search.Foundations and Trends in Machine Learning,2014,7(1):1–130.[22]R.R.Negenborn,B.De Schutter,H.Hellendoorn.Multi-agent model predictive control for transportation networks:Serial versus parallel schemes.Engineering Applications of Artificial Intelligence,2008,21(3):353–366.

    [23]J.Liu,X.Chen,D.M.de la Pe?na,et al.Sequential and iterative architectures for distributed model predictive control of nonlinear process systems.American Institute of Chemical Engineers(AIChE)Journal,2010,56(8):2137–2149.

    [24]L.Bu?soniu,R.Munos,Robert Babuˇska.A review of optimistic planning in Markov decision processes.Reinforcement Learning and Adaptive Dynamic Programming for Feedback Control.F.Lewis,D.Liu(eds.),Hoboken:John Wileyamp;Sons,2012:DOI 10.1002/9781118453988.ch22.

    [25]L.Bu?soniu,D.Ernst,B.De Schutter,et al.Approximate dynamic programming with a fuzzy parameterization.Automatica,2010,46(5):804–814.

    [26]B.Kiumarsi,F.Lewis,H.Modares,et al.Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics.Automatica,2014,50(4):1167–1175.

    the M.Sc.degree(valedictorian)from the Technical University of Cluj-Napoca,Cluj-Napoca,Romania,in 2003,and the Ph.D.degree(cum laude)from the Delft University of Technology,Delft,the Netherlands,in 2009.He is an associate professor with the Department of Automation,Technical University of Cluj-Napoca,Romania.His research interests include planning-based methods for nonlinear optimal control,reinforcement learning and dynamic programming with function approximation,multi-agent systems,and,more generally,intelligent and learning techniques for control.He has authored a book as well as a number of journals,conferences,and chapter publications on these topics.Dr.Bu?soniu was the recipient of the 2009 Andrew P.Sage Award for the best paper in the IEEE Transactions on Systems,Man,and Cybernetics.E-mail:lucian@busoniu.net.

    Irinel-ConstantinMOR?ARESCUis currently an associate professor at Universit′e de Lorraine and a researcher at the Research Centre of Automatic Control(CRAN UMR 7039 CNRS)in Nancy,France.He received the B.Sc.and the M.Sc.degrees in Mathematics from University of Bucharest,Romania,in 1997 and 1999,respectively.In 2006,he received the Ph.D.degree in Mathematics and Technology of Information and Systems from University of Bucharest and University of Technology of Compi`egne,respectively.His works concern stability and controloftime-delay systems,tracking control for nonsmooth mechanical systems,consensus and synchronization problems.E-mail:constantin.morarescu@univ-lorraine.fr.

    ?Corresponding author.

    E-mail:lucian@busoniu.net.

    This work was supported by a Programme Hubert Curien-Brancusi cooperation grant(CNCS-UEFISCDI contract no.781/2014 and Campus France grant no.32610SE).Additionally,the work of L.Bu?soniu was supported by the Romanian National Authority for Scientific Research,CNCS-UEFISCDI(No.PNII-RU-TE-2012-3-0040).The work of I.-C.Mor?arescu was partially funded by the National Research Agency(ANR)project“Computation Aware Control Systemsquot;(No.ANR-13-BS03-004-02).

    ?2015 South China University of Technology,Academy of Mathematics and Systems Science,CAS,and Springer-Verlag Berlin Heidelberg

    亚洲无线在线观看| 国产成年人精品一区二区| 欧美三级亚洲精品| 搞女人的毛片| 最新美女视频免费是黄的| 国产成人欧美在线观看| 麻豆成人午夜福利视频| 高清毛片免费观看视频网站| 亚洲性夜色夜夜综合| 久久久久久国产a免费观看| 亚洲自拍偷在线| 在线观看66精品国产| 欧美乱色亚洲激情| 中文字幕久久专区| 免费看日本二区| 亚洲一卡2卡3卡4卡5卡精品中文| 丝袜美腿诱惑在线| 宅男免费午夜| 亚洲真实伦在线观看| 精品电影一区二区在线| 国产精品1区2区在线观看.| 国产精品99久久99久久久不卡| 欧美 亚洲 国产 日韩一| 9191精品国产免费久久| 中文字幕久久专区| 久久国产精品人妻蜜桃| 国产v大片淫在线免费观看| 好男人在线观看高清免费视频| 免费搜索国产男女视频| 亚洲av成人av| 国产主播在线观看一区二区| 午夜免费成人在线视频| 在线观看www视频免费| 日本成人三级电影网站| 婷婷精品国产亚洲av| 好看av亚洲va欧美ⅴa在| 两个人看的免费小视频| 日韩欧美国产在线观看| 国产精品美女特级片免费视频播放器 | 亚洲精品粉嫩美女一区| 久久精品aⅴ一区二区三区四区| 精华霜和精华液先用哪个| 悠悠久久av| 999精品在线视频| 久久国产精品影院| 国产精品亚洲美女久久久| 一级毛片精品| 黄色a级毛片大全视频| 高清在线国产一区| 婷婷丁香在线五月| 亚洲成人久久爱视频| 欧美日韩一级在线毛片| 18禁裸乳无遮挡免费网站照片| 黄频高清免费视频| 黄色视频,在线免费观看| 亚洲精品色激情综合| 日本精品一区二区三区蜜桃| 亚洲,欧美精品.| 中文字幕人妻丝袜一区二区| 久久精品国产亚洲av香蕉五月| 在线视频色国产色| 精品电影一区二区在线| 久久久国产欧美日韩av| 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲中文日韩欧美视频| 国产99久久九九免费精品| 丰满人妻熟妇乱又伦精品不卡| 久久久久国产一级毛片高清牌| 三级毛片av免费| 日韩有码中文字幕| 久久九九热精品免费| 午夜福利视频1000在线观看| 少妇裸体淫交视频免费看高清 | 国产精品亚洲av一区麻豆| 桃色一区二区三区在线观看| 亚洲人成伊人成综合网2020| 18禁黄网站禁片午夜丰满| 桃红色精品国产亚洲av| 免费看十八禁软件| 1024视频免费在线观看| 这个男人来自地球电影免费观看| 桃红色精品国产亚洲av| 国产欧美日韩一区二区精品| 欧美日韩一级在线毛片| 欧美色欧美亚洲另类二区| 久久久精品国产亚洲av高清涩受| 黄色片一级片一级黄色片| 国产亚洲av嫩草精品影院| 男人的好看免费观看在线视频 | 成年版毛片免费区| 日韩国内少妇激情av| 欧美性长视频在线观看| 日韩精品中文字幕看吧| 国产精品1区2区在线观看.| 动漫黄色视频在线观看| 国产高清视频在线播放一区| 国产精品综合久久久久久久免费| 成人国产一区最新在线观看| 老熟妇仑乱视频hdxx| 亚洲精品在线美女| 精品国产亚洲在线| 色在线成人网| 亚洲av成人不卡在线观看播放网| 亚洲专区中文字幕在线| 97碰自拍视频| 日本五十路高清| 日韩欧美精品v在线| 在线视频色国产色| 国产一区在线观看成人免费| 国产精品久久久av美女十八| 一级a爱片免费观看的视频| www日本黄色视频网| 国产精品一区二区三区四区久久| 亚洲精品粉嫩美女一区| 九九热线精品视视频播放| 久久久精品欧美日韩精品| 国产伦人伦偷精品视频| 久久天躁狠狠躁夜夜2o2o| 毛片女人毛片| x7x7x7水蜜桃| 性欧美人与动物交配| 国产精品 欧美亚洲| 亚洲国产欧美网| 中文字幕精品亚洲无线码一区| 亚洲色图av天堂| 免费看日本二区| 中文资源天堂在线| 国产一区二区在线av高清观看| 国产在线观看jvid| av天堂在线播放| 黄色女人牲交| 天堂动漫精品| 十八禁网站免费在线| 久久人妻福利社区极品人妻图片| 欧美在线一区亚洲| 亚洲色图 男人天堂 中文字幕| 久久这里只有精品19| 99热这里只有精品一区 | 国内毛片毛片毛片毛片毛片| 国产高清有码在线观看视频 | 又爽又黄无遮挡网站| 久久久久性生活片| 亚洲avbb在线观看| 一级毛片精品| 熟妇人妻久久中文字幕3abv| 国产精品99久久99久久久不卡| 亚洲精品在线美女| 国产精品久久久久久亚洲av鲁大| 日韩成人在线观看一区二区三区| 国内揄拍国产精品人妻在线| 国产高清视频在线观看网站| 久久久久久大精品| 国产精品久久久久久人妻精品电影| 一个人观看的视频www高清免费观看 | 99国产精品99久久久久| 人妻丰满熟妇av一区二区三区| svipshipincom国产片| 亚洲第一电影网av| 亚洲avbb在线观看| 亚洲在线自拍视频| 国产亚洲精品综合一区在线观看 | 国产午夜精品久久久久久| 欧美黑人欧美精品刺激| 国产伦人伦偷精品视频| 久久午夜亚洲精品久久| 97超级碰碰碰精品色视频在线观看| 精品欧美一区二区三区在线| 欧美国产日韩亚洲一区| 午夜日韩欧美国产| 精品不卡国产一区二区三区| 国产男靠女视频免费网站| 色噜噜av男人的天堂激情| 日韩欧美三级三区| 88av欧美| 一个人免费在线观看电影 | 日本 欧美在线| 18禁黄网站禁片午夜丰满| 黄色丝袜av网址大全| 精品高清国产在线一区| www日本在线高清视频| 久久精品91蜜桃| 亚洲乱码一区二区免费版| 午夜免费观看网址| 19禁男女啪啪无遮挡网站| 亚洲18禁久久av| 久久天躁狠狠躁夜夜2o2o| 97超级碰碰碰精品色视频在线观看| 天堂影院成人在线观看| 在线国产一区二区在线| 国产99白浆流出| 国产真人三级小视频在线观看| 亚洲精品粉嫩美女一区| 成人午夜高清在线视频| 老司机午夜十八禁免费视频| 黄色丝袜av网址大全| 看黄色毛片网站| 不卡一级毛片| 国产一区二区在线观看日韩 | 国产精品自产拍在线观看55亚洲| 少妇裸体淫交视频免费看高清 | 国产片内射在线| 亚洲天堂国产精品一区在线| 国产精品野战在线观看| 国产乱人伦免费视频| 男人舔奶头视频| 99久久精品热视频| 国产精品自产拍在线观看55亚洲| 母亲3免费完整高清在线观看| 高清毛片免费观看视频网站| 热99re8久久精品国产| 精品第一国产精品| 无限看片的www在线观看| 午夜视频精品福利| 久久久久久久久久黄片| 人成视频在线观看免费观看| 99久久国产精品久久久| 高清在线国产一区| 好看av亚洲va欧美ⅴa在| 欧美一区二区精品小视频在线| 他把我摸到了高潮在线观看| 国产黄a三级三级三级人| 国产精品久久视频播放| xxxwww97欧美| 怎么达到女性高潮| 欧美成人午夜精品| 国产成人系列免费观看| 久久精品综合一区二区三区| 久久久久久久精品吃奶| 午夜激情福利司机影院| 精品久久久久久久久久久久久| www日本在线高清视频| 午夜影院日韩av| 五月玫瑰六月丁香| 最近最新免费中文字幕在线| 免费看日本二区| 国产熟女xx| 成人国产一区最新在线观看| 国产私拍福利视频在线观看| 嫩草影院精品99| 国产av又大| 日本一本二区三区精品| 黄色成人免费大全| 日韩成人在线观看一区二区三区| a级毛片a级免费在线| 在线观看日韩欧美| 制服人妻中文乱码| 国产精品国产高清国产av| 日韩欧美国产一区二区入口| 老汉色av国产亚洲站长工具| 毛片女人毛片| 久久午夜综合久久蜜桃| 国产精品一区二区精品视频观看| 欧美av亚洲av综合av国产av| 最近视频中文字幕2019在线8| 一进一出抽搐动态| 正在播放国产对白刺激| 国产精品乱码一区二三区的特点| 欧美又色又爽又黄视频| 久久久国产精品麻豆| 国产精华一区二区三区| 国产久久久一区二区三区| 国产精品乱码一区二三区的特点| 久久人人精品亚洲av| 叶爱在线成人免费视频播放| 三级毛片av免费| 精品高清国产在线一区| 免费观看精品视频网站| 变态另类成人亚洲欧美熟女| 国产成人系列免费观看| 我要搜黄色片| 国产日本99.免费观看| 看片在线看免费视频| 日本黄色视频三级网站网址| 亚洲片人在线观看| 日本免费a在线| bbb黄色大片| 国内精品久久久久精免费| 又紧又爽又黄一区二区| 99re在线观看精品视频| 日韩欧美免费精品| 免费在线观看成人毛片| 看免费av毛片| 国产亚洲精品久久久久久毛片| av欧美777| 日本免费a在线| 亚洲欧美精品综合久久99| 国产亚洲精品综合一区在线观看 | 日本一二三区视频观看| 一本精品99久久精品77| 妹子高潮喷水视频| 日本精品一区二区三区蜜桃| 日韩大码丰满熟妇| 黄色丝袜av网址大全| 变态另类成人亚洲欧美熟女| 久久这里只有精品中国| 黄色a级毛片大全视频| 观看免费一级毛片| 一二三四在线观看免费中文在| 亚洲自拍偷在线| 久久香蕉精品热| 久久精品国产亚洲av香蕉五月| 精品久久久久久久末码| 这个男人来自地球电影免费观看| 一卡2卡三卡四卡精品乱码亚洲| 搞女人的毛片| 男人舔女人下体高潮全视频| 制服诱惑二区| 久久久久久久久免费视频了| 久久精品91蜜桃| 亚洲激情在线av| 黄片大片在线免费观看| 国产aⅴ精品一区二区三区波| 熟女电影av网| 国产91精品成人一区二区三区| 久久精品成人免费网站| 亚洲在线自拍视频| 亚洲一区二区三区色噜噜| 亚洲av成人av| 久久香蕉激情| 精品高清国产在线一区| 男人舔女人的私密视频| 黄色视频不卡| 亚洲成a人片在线一区二区| 中文字幕av在线有码专区| 天天躁夜夜躁狠狠躁躁| 欧美一区二区国产精品久久精品 | 真人做人爱边吃奶动态| 成人国产一区最新在线观看| 听说在线观看完整版免费高清| 又爽又黄无遮挡网站| 国产精品影院久久| 亚洲精品粉嫩美女一区| x7x7x7水蜜桃| 成人高潮视频无遮挡免费网站| 午夜a级毛片| 亚洲五月婷婷丁香| 91老司机精品| 精品人妻1区二区| 最近视频中文字幕2019在线8| 久久久久国产一级毛片高清牌| 欧美日韩瑟瑟在线播放| 国产一区二区在线观看日韩 | 午夜精品一区二区三区免费看| 99国产极品粉嫩在线观看| 香蕉丝袜av| 国内久久婷婷六月综合欲色啪| 国产精品亚洲一级av第二区| 岛国视频午夜一区免费看| 无人区码免费观看不卡| 搡老熟女国产l中国老女人| 午夜福利18| 成人午夜高清在线视频| 丰满人妻熟妇乱又伦精品不卡| 国产99白浆流出| 久久久久久久午夜电影| www国产在线视频色| 国产精品99久久99久久久不卡| 18禁国产床啪视频网站| 日韩大尺度精品在线看网址| 不卡av一区二区三区| 色综合亚洲欧美另类图片| 最近在线观看免费完整版| 女人爽到高潮嗷嗷叫在线视频| 一区福利在线观看| 麻豆av在线久日| 国内精品久久久久精免费| 欧美+亚洲+日韩+国产| 丝袜美腿诱惑在线| 成人18禁在线播放| 亚洲精品中文字幕在线视频| 天堂√8在线中文| 桃色一区二区三区在线观看| 日韩国内少妇激情av| 亚洲18禁久久av| 非洲黑人性xxxx精品又粗又长| 成人国产综合亚洲| 国产精品久久久人人做人人爽| 最近在线观看免费完整版| 亚洲自偷自拍图片 自拍| 日韩精品免费视频一区二区三区| 在线观看午夜福利视频| a级毛片a级免费在线| 首页视频小说图片口味搜索| 久久国产精品影院| 男女床上黄色一级片免费看| 在线观看日韩欧美| 亚洲va日本ⅴa欧美va伊人久久| 在线永久观看黄色视频| 午夜精品在线福利| 99久久99久久久精品蜜桃| av在线播放免费不卡| 99riav亚洲国产免费| 国产在线精品亚洲第一网站| av天堂在线播放| 亚洲精华国产精华精| 一级毛片精品| 亚洲成人中文字幕在线播放| 国产精品 国内视频| 国产av麻豆久久久久久久| 女人高潮潮喷娇喘18禁视频| 最近最新免费中文字幕在线| 看片在线看免费视频| 国产不卡一卡二| 久99久视频精品免费| 老熟妇乱子伦视频在线观看| 久久99热这里只有精品18| 日本一区二区免费在线视频| 亚洲色图av天堂| 久久精品影院6| 精品午夜福利视频在线观看一区| 精品国产亚洲在线| 一个人免费在线观看的高清视频| 亚洲av电影不卡..在线观看| 日本黄色视频三级网站网址| 亚洲无线在线观看| 色综合欧美亚洲国产小说| 日本免费一区二区三区高清不卡| 久久人妻福利社区极品人妻图片| 欧美丝袜亚洲另类 | 麻豆久久精品国产亚洲av| 丝袜美腿诱惑在线| 正在播放国产对白刺激| 国产爱豆传媒在线观看 | 国产亚洲精品综合一区在线观看 | 首页视频小说图片口味搜索| 夜夜看夜夜爽夜夜摸| a级毛片a级免费在线| 国产高清videossex| 久久久国产欧美日韩av| 男插女下体视频免费在线播放| 天天一区二区日本电影三级| 亚洲国产欧洲综合997久久,| 狂野欧美激情性xxxx| 麻豆久久精品国产亚洲av| 大型av网站在线播放| 亚洲国产欧美网| 免费av毛片视频| 国产av又大| 欧美高清成人免费视频www| 国产精品香港三级国产av潘金莲| 黄色成人免费大全| 脱女人内裤的视频| av在线播放免费不卡| 国产成人aa在线观看| 国产精品免费视频内射| 91成年电影在线观看| 最近最新中文字幕大全电影3| 国产三级在线视频| 香蕉国产在线看| 欧美精品啪啪一区二区三区| 亚洲熟妇中文字幕五十中出| 成熟少妇高潮喷水视频| www.自偷自拍.com| a级毛片在线看网站| 91九色精品人成在线观看| 岛国视频午夜一区免费看| 人妻久久中文字幕网| 亚洲专区中文字幕在线| 99riav亚洲国产免费| 熟妇人妻久久中文字幕3abv| 99在线人妻在线中文字幕| 亚洲av电影不卡..在线观看| 波多野结衣高清无吗| 香蕉国产在线看| 欧美日韩精品网址| 免费观看人在逋| 亚洲av成人精品一区久久| 国产一区在线观看成人免费| 在线观看日韩欧美| 精品不卡国产一区二区三区| 免费观看人在逋| 国产1区2区3区精品| 香蕉av资源在线| 成熟少妇高潮喷水视频| 日本五十路高清| 在线观看免费日韩欧美大片| 亚洲自偷自拍图片 自拍| 欧美一区二区国产精品久久精品 | 国产伦一二天堂av在线观看| 亚洲精品美女久久av网站| 丰满人妻熟妇乱又伦精品不卡| 成人三级黄色视频| 国产成人aa在线观看| 国产精品久久久久久久电影 | 一二三四在线观看免费中文在| 少妇裸体淫交视频免费看高清 | av福利片在线观看| 精品久久久久久久毛片微露脸| 久久午夜综合久久蜜桃| 日韩精品免费视频一区二区三区| 99久久精品国产亚洲精品| 午夜两性在线视频| 欧美zozozo另类| 亚洲va日本ⅴa欧美va伊人久久| videosex国产| 久久人妻av系列| 国产高清激情床上av| 欧美3d第一页| 久久久久亚洲av毛片大全| 制服诱惑二区| 一二三四在线观看免费中文在| 国产亚洲精品久久久久5区| 啦啦啦观看免费观看视频高清| 18禁国产床啪视频网站| www.精华液| 美女免费视频网站| 亚洲电影在线观看av| 狠狠狠狠99中文字幕| 亚洲国产日韩欧美精品在线观看 | 观看免费一级毛片| 国产av一区二区精品久久| 婷婷丁香在线五月| 午夜福利高清视频| 俺也久久电影网| 国语自产精品视频在线第100页| 极品教师在线免费播放| 久久国产精品人妻蜜桃| 久久精品国产亚洲av香蕉五月| 欧美一级毛片孕妇| 久久久国产欧美日韩av| 精品无人区乱码1区二区| 欧美+亚洲+日韩+国产| 久久香蕉激情| 亚洲 欧美 日韩 在线 免费| 久久天躁狠狠躁夜夜2o2o| 在线十欧美十亚洲十日本专区| 99国产综合亚洲精品| 日日爽夜夜爽网站| 国产高清视频在线观看网站| 国产在线精品亚洲第一网站| 一进一出好大好爽视频| 日本在线视频免费播放| 99re在线观看精品视频| 亚洲第一欧美日韩一区二区三区| 婷婷精品国产亚洲av在线| 在线观看免费午夜福利视频| 国产一级毛片七仙女欲春2| 国产成人影院久久av| 久热爱精品视频在线9| 日韩av在线大香蕉| 超碰成人久久| av在线天堂中文字幕| 国产亚洲精品综合一区在线观看 | 午夜久久久久精精品| 久久婷婷人人爽人人干人人爱| www.精华液| 日本 av在线| 蜜桃久久精品国产亚洲av| 亚洲精品久久成人aⅴ小说| 午夜成年电影在线免费观看| 亚洲成人中文字幕在线播放| 国内精品久久久久久久电影| 久久欧美精品欧美久久欧美| bbb黄色大片| 日韩免费av在线播放| 午夜精品久久久久久毛片777| 亚洲精品在线美女| 一本一本综合久久| 亚洲一区二区三区色噜噜| 日韩高清综合在线| 国产乱人伦免费视频| 黄色视频,在线免费观看| 男女那种视频在线观看| 国产在线观看jvid| 久久久国产成人免费| 激情在线观看视频在线高清| АⅤ资源中文在线天堂| 首页视频小说图片口味搜索| 亚洲人成网站在线播放欧美日韩| 中文字幕熟女人妻在线| www.熟女人妻精品国产| av国产免费在线观看| 在线观看免费午夜福利视频| 999精品在线视频| 女警被强在线播放| 日韩精品青青久久久久久| 国产成人一区二区三区免费视频网站| 亚洲真实伦在线观看| 色综合欧美亚洲国产小说| 亚洲精品久久成人aⅴ小说| 一卡2卡三卡四卡精品乱码亚洲| 婷婷六月久久综合丁香| www.熟女人妻精品国产| 久久中文看片网| 怎么达到女性高潮| 国产精品日韩av在线免费观看| 9191精品国产免费久久| 搡老妇女老女人老熟妇| 老司机在亚洲福利影院| 十八禁人妻一区二区| 老司机靠b影院| 久久这里只有精品19| 亚洲美女黄片视频| 人妻久久中文字幕网| 久久天堂一区二区三区四区| 人妻丰满熟妇av一区二区三区| 久久久国产精品麻豆| 欧美日韩国产亚洲二区| 性色av乱码一区二区三区2| 色老头精品视频在线观看| 国模一区二区三区四区视频 | av福利片在线观看| 午夜福利视频1000在线观看| 国产爱豆传媒在线观看 | 波多野结衣高清无吗| 久久久久久亚洲精品国产蜜桃av| 18禁黄网站禁片免费观看直播| 亚洲欧洲精品一区二区精品久久久| 久久久久久久久久黄片| 窝窝影院91人妻| 九色成人免费人妻av| 黑人操中国人逼视频| 狠狠狠狠99中文字幕| 黑人巨大精品欧美一区二区mp4| 国产av又大|