• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep Reinforcement Learning-Based Task Offloading and Service Migrating Policies in Service Caching-Assisted Mobile Edge Computing

    2024-04-28 11:59:08KeHongchangWangHuiSunHongbinHalvinYang
    China Communications 2024年4期

    Ke Hongchang ,Wang Hui ,Sun Hongbin ,Halvin Yang

    1 School of Computer Technology and Engineering,Changchun Institute of Technology,Changchun 130012,China

    2 College of Computer Science and Engineering,Changchun University of Technology,Changchun 130012,China

    3 Department of Electronic and Electrical Engineering,University College London,London,UK

    Abstract: Emerging mobile edge computing(MEC)is considered a feasible solution for offloading the computation-intensive request tasks generated from mobile wireless equipment(MWE)with limited computational resources and energy.Due to the homogeneity of request tasks from one MWE during a longterm time period,it is vital to predeploy the particular service cachings required by the request tasks at the MEC server.In this paper,we model a service caching-assisted MEC framework that takes into account the constraint on the number of service cachings hosted by each edge server and the migration of request tasks from the current edge server to another edge server with service caching required by tasks.Furthermore,we propose a multiagent deep reinforcement learning-based computation offloading and task migrating decision-making scheme(MBOMS)to minimize the long-term average weighted cost.The proposed MBOMS can learn the near-optimal offloading and migrating decision-making policy by centralized training and decentralized execution.Systematic and comprehensive simulation results reveal that our proposed MBOMS can converge well after training and outperforms the other fvie baseline algorithms.

    Keywords: deep reinforcement learning;mobile edge computing;service caching;service migrating

    I.INTRODUCTION

    In heterogeneous wireless networks,mobile wireless equipment (MWE) generates a quantity of computation-intensive tasks(such as unmanned vehicle maps,somatosensory games,virtual reality/augmented reality(VR/AR)),which cannot be processed locally by their own processor due to limited resources[1,2].Therefore,how to process and complete these request tasks within a stringent latency constraint for management of heterogeneous networks is a critical issue,such as ameliorating the quality of service(QoS),minimizing the total cost or maximizing the overall utility[3].

    Emerging mobile edge computing (MEC) technology is a promising paradigm for transmitting the request tasks of MWEs to edge servers by uplink wireless channels.In addition,the rich computational resources of the MEC server can meet the computational demands of request tasks from MWEs [4,5].The proximity of the MEC server to MWEs makes the transmission tasks extremely convenient and backhaul latency low[6,7].The request tasks can be easily offloaded to the MEC server with powerful computation ability,which can not only improve QoS but also strengthen MWE’s adaptability for computationintensive or latency-sensitive applications [8].There have been many works on computational offloading or resource allocation in the MEC environment.Liuet al.formulated a method for job scheduling of MEC resources in multiuser scenes.The method could utilize the improved heuristic algorithm (binary-relaxed solution) to make the job placement decision to minimize the long-term slack time of users[9].Shenget al.presents a computation offloading decision-making policy in the MEC scenario to minimize the average latency cost of all users.The mentioned scheme considered power allocation,resource allocation and computational offloading based on optimization theory and matching theory[10].However,in[9,10],the works assume that the computational resources of the MEC server are suffciient for multiuser scenes.In fact,although the resources of MEC servers are abundant,the computation-intensive tasks of massive users will overwhelm MEC servers.

    Because of the homogeneity of request tasks from one MWE during a long-term time period,it is vital to predeploy the particular cachings required by the request tasks at the MEC server.Caching can be typed as content caching and service caching.Content caching is the static content (such as video and gaming)in the content buffer of the MEC server[11,12].However,service caching can be viewed as caching of dynamic content.The dynamic content here is the executable program,which is called computation realized by the program such as required libraries,databases,or computation resources for executing the request tasks.In terms of content caching in the MEC environment,the static characteristic of that is considered,such as Sunet al.proposed a joint optimization scheme for caching and computing VR in an MEC system [13],HUYNHet al.implemented an edge caching MEC network to fnid the optimal decisionmaking policy.There are a bit existing works on service caching.Liet al.introduced a computational offloading decision-making policy in a service cachingassisted MEC network using game theory to obtain the minimum power consumption cost under the time constraint[14].Chenet al.formulated a service cachingassisted offloading optimization scheme to minimize the long-term average task fniished delay[15].However,the aforementioned works in[14,15]did not consider the time-varying channel state between mobile users and MEC servers or cloud servers and resource sharing among MEC servers.In addition,game theory and the ADMM optimization method consider individuals more than resource competition.

    Since the introduction of deep reinforcement learning (DRL),signifciant results have been achieved in both theory and application in artifciial intelligence[16].There are several DRL algorithms composed of value-based methods(deep Q-Learning(DQN)[17],double DQN (DDQN)[18],and dueling DQN[19]),policy-based methods(policy gradient(PG)[20],proximal policy optimization (PPO)[21]),and combining policy and value methods (actor-critic) [22] and deep deterministic policy gradient(DDPG)[23]).There are many works on DRL-based computation offloading or task migrating in MEC frameworks,such as Liuet al.designed a multiuser system and proposed a DQNbased offloading optimization policy[24],Elgendyet al.utilized the Q-Learning and DQN algorithms to tackle the joint offloading and caching optimization problem [25].Wanget al.presented a joint computation migration and resource allocation policy based on DDQN in vehicular networks[26].However,in[24-26],the mentioned DRL-based schemes are single agent or arbitrary decentralized multiagent learning patterns that are not suited for the massive MWE and MEC server scenarios in heterogeneous MEC networks.

    Deep multiagent reinforcement learning (MARL)algorithms can train neural networks in centralized or decentralized patterns,which have been applied in MEC environments in some existing works.In terms of state-of-the-art MARL-based scheme,Gonget al.designed a joint optimization scheme based on the multiagent deep deterministic policy gradient(MADDPG)for fniding near-optimal offloading,computational resources and power allocation decisionmaking policy to minimize the total overhead composed of delay cost and energy consumption cost[27].Seidet al.formulated the joint optimization for offloading,computation allocation and association problem to stochastic game (SG) and used the MADDPG algorithm with centralized training and decentralized execution characteristic to learn the nearoptimal decision-making policy for obtaining the minimum computation cost of total MEC system[28].Although the aforementioned works in [27,28] implemented the MARL-based solution for computationintensive tasks from multiple users in a multi-edge server environment,they did not take into account the interaction among different MEC servers and the timevarying achievable capacity of MEC servers.However,there were also some state-of-the-art schemes related to our work.Houet al.transited the offloading optimal problem into Partially Observable Markov Decision Process(POMDP)and constructed the deep neural network based on Double Deep Q Network(DDQN)to learn the optimal offloading decision[29].The designed MEC environment of[29]was more theoretical,without considering the randomness of request tasks generated by MWEs and the time-varying characteristic of channel state.Sunet al.formulated a multi-user and multi-cell MEC framework and proposed a multi-branch network DDQN-based computation offloading and resource allocation decisionmaking policy to obtain the minimum task delay and discard rate [30].Although the work in [30] implemented offloading decisions for multi user terminals in a multi edge server environment,it did not considered cooperative scheduling between edge servers.Once the resource of an edge server was insuffciient or did not have the required service caching for specifci tasks,it will affect effciiency of algorithm.

    To address aforementioned challenges,in this paper,we design a multi-service caching-enabled MEC network and propose an MADDPG-based computation offloading and task migrating decision-making scheme(MBOMS)to fnid the optimal policy for minimizing the long-term average weighted cost,including the latency cost and energy consumption cost.To the best of our knowledge,our work is the frist to jointly consider the randomness of the request tasks generated by the MWE,the time-varying channel between the MWE and the edge server,the changing capacity and hosted service caching of the edge server,and the different priorities of request tasks.

    The main contributions of our work are illustrated as follows:

    ? To effectively tackle the computation-intensive request tasks generated by the resource-limited MWEs,we design a service caching-assisted MEC framework that takes into account the constraint on the number of service cachings hosted by each edge server and the migration of request tasks from the current edge server to another edge server with service caching required by tasks.

    ? By considering a multiuser MEC network model with a time-varying channel state,the changing capacity of the edge server,and the different priorities of request tasks,we formulate the joint offloading and migrating problem as an MDP to minimize the long-term average weighted cost,including the latency cost and energy consumption cost.

    ? Because of noprioriknowledge of state space and state transition probabilities in each time slot,we transit the MDP-based scheme to the MADDPGbased problem.Each MWE can act as a learning agent and interact with the MEC environment.By offline training of the neural networks of the MADDPG model,the near-optimal offloading and migrating policy can be learned with centralized training and decentralized execution.

    ? To evaluate and verify the convergence and effectiveness of the proposed MBOMS,we perform systematic and comprehensive numerical simulations.Different parameter confgiurations for MBOMS in the simulation results reveal good convergence,and a comparison with fvie baseline algorithms indicates the robustness and superiority of MBOMS.

    The rest of this paper is organized as follows.A service caching-enabled MEC network model is introduced in Section II.Section III illustrates the joint offloading and migrating scheme for minimizing the average weighted cost of the overall MEC system.In Section IV,an MADDPG-based optimization scheme for offloading and migrating decisions is proposed.Section V provides the systematic numerical simulation results.Our work is concluded in Section VI.

    II.SYSTEM MODEL

    To better describe the system model,in this section,we consider a service caching-assisted MEC system composed of the network model,the communication model,and the computation model.

    2.1 Network Model

    We designed a three-tier architecture system assisted with service caching,including the mobile user layer,the MEC layer,and the cloud layer.Figure 1 shows the edge server framework assisted by service caching.We denoteE={1,2,···,E},R={1,2,···,R},andC={1,2,···,C}as the set of MWEs,request tasks,and service cachings,respectively.Each request task from MWEs requires certain service caching for computing.We assume thatre,c,re,c ∈Ris a request task generated by MWEe,which needs service cachingc.N={1,2,···,N}stands for the set of edge servers.M={1,2,···,M}andU={1,2,···,U}are the sets of MEC servers and UAVs,respectively.We can see thatN=M ∪U,andM ∩U=?.Furthermore,we assume that if edge servern(MEC servermor UAVu) hosts the server c achingc,taskre,ccan be executed by edge servern;otherwise,it cannot,whereis the subset ofC.

    Figure 1.Service caching-assisted MEC framework.

    Since each edge server cannot host all cachings,if the edge server does have the service caching required by the request task,these tasks can be migrated to the edge server with the corresponding service caching for processing.If all server edges do not host the server caching required for the request task,the task can be offloaded to the cloud server for processing.To this end,we defnieωr,ωr,nas the migrating decision and migrate to which edge server for taskre,c,meetingωr ∈[0,1],ωr,n ∈[0,1,...,N].That is,if taskre,con edge servernneeds to be migrated to another edge server or offloaded to the cloud server,ωr=1;otherwise,ωr=0.If taskre,con edge servernmigrates to edge serverand if taskre,con edge servernis offloaded to the cloud server,ωr,n=0.Table 1 lists the main parameters and descriptions in our work.

    Table 1.List of main notations.

    2.2 Communication Model

    We consider discretizing the continuous time period to the time slot (certain time intervals),which is defnied asT={1,2,···,T}.The channels between the MWE and the edge server,the edge server and the cloud server are time-varying in different time slots,but they are the same in each time slot.We assume that an MWE can only communicate with the edge server(an MEC server or a UAV) closest to it regardless ofthe remaining computing resources and hosted service cachings of the edge server in each time slot.That is,if the MWE generates the request tasks,the tasks are frist offloaded to the edge server closest.Then,the tasks are migrated to the other edge server or offloaded to the cloud server according to the service cachings hosted by which edge server.We defnie the channel state between MWEEand the edge servernin time slottas

    whereρe,npresents the path-loss coeffciient.αe,nis the constant for controlling the scale of the channel state.Schannel(t) is the channel state in current time slot.de,n(t)stands for the distance between MWEEand the edge servern.

    In our paper,the channel states are time-varying in different time slots and modeled as a Markov distribution,meetingAn,c=P(Schannel(t+1)=pk|Schannel(t)=pl),whereAn,cis the channel state transition probability matrix,pk,pl ∈Schannel,andSchannelis the channel state vector.

    The maximum offloading transmission rate between the MWEeand the edge servernby Shannon’s law can be computed by

    whereζnis the ratio of bandwidth allocated to MWEefor offloading the request tasks andWnstands for the total bandwidth of edge servern.pe,n(t) represents the offloading transmission power for tasks from MWEeto edge servern.σn(t)denotes the variance of the additive white Gaussian noise(AWGN)with zero mean.

    2.3 Computation Model

    As shown in Section 2.1,a certain request task requires particular server caching,which can greatly alleviate the computational pressure on the edge server;for example,the real-time scene map for unmanned vehicles needs the particular server caching located at the edge server or cloud server.To this end,the request task can be offloaded to the edge server with server caching for computation by wireless communication.We defnie the request tasksre,c(t) in time slottas the tuple,meetingwhereis the total size of request tasksre,c(t),stands for the computation density(number of required CPU cycles) for request tasksre,c(t),andrepresents the constraint on completion time(the deadline for completing request tasksre,c(t)).is the priority of completing request taskre,c(t).Furthermore,we defniePr(t)as the arrival probability of request tasksrfrom MWEein time slottfollowing a Bernoulli distribution withνe.Thereafter,the arrival probability meets E[Pr(I(re,c(t)=1))]=νe,whereI(.) is the indicator function,which means that if MWEegenerates the request tasksre,c(t),I(re,c)=1;otherwise,I(re,c)=0.

    2.3.1 Computation Offloading to Edge Server

    We defnieBn,qas the task buffer queue for storing the different priority request tasks from the linked MWE.Therefore,the transmission latency for offloading the tasks from MWEeto edge serverncan be calculated by

    Accordingly,the energy consumption for transmitting the offloading tasks from MWEeto edge serverncan be calculated by

    If edge servernhosts service cachingcfor request taskre,coffloaded by NWEe,edge serverncan execute request taskre,cby its own allocation computation resource with service cachingc,and the execution latency of computing request taskre,cis derived by

    wherefe,c(t)is the total number of allocated CPU cycles by edge servernfor computing request taskre,c.In this paper,fe,c(t) is related toand can be defnied as

    To this end,the executed energy consumption of computing the request taskre,cis[33]

    whereκnstands for the achievable switched parameter based on chip architecture characteristics of the edge servern.

    Because one edge server needs to process multiple request tasks,we defnie the task buffer queues to relieve the computing pressure of the edge server.Therefore,request tasks are frist stored in the buffer queue,so the total latency of completing tasks includes the waiting latency,which is computed as

    whereξn,qis the weighted coeffciient,which is related to the priorityof request tasks,and according to Little’s Law,waiting latency is proportional to the total length of buffer queueBn,q.

    The waiting energy consumption can be defnied as

    2.3.2 Migration Tasks to the Other Edge Server

    Once the edge serverndoes not host the service cachingcrequired by the offloaded request tasksre,cfrom the MWEe,request tasksre,ccannot be executed by the edge servern.Therefore,the request tasks may be migrated to the other edge server if it hosts service cachingc.Under these circumstances,we should consider the migration latency,which can be written as

    Accordingly,the energy consumption for task migration from edge servernto edge serveris

    2.3.3 Computation Offloading to Cloud Server

    If all edge servers do not host the service cachingcrequired by the offloaded request tasksre,cfrom the MWEe,the tasksre,cneed to be offloaded from the edge server to the cloud server for execution.Therefore,the transmission latency for offloading tasks from the edge servernto the cloud server can be written as

    To this end,the energy consumption for transmitting the offloading tasks from edge servernto the cloud server is derived by

    Compared with edge servers,cloud servers have suffciient computational resources and service cachings,and the execution latency,energy consumption,waiting latency,and energy consumption can be ignored.However,we consider the downlink transmission latency and energy consumption for the computation result from the cloud server to the edge servern,which are defnied as

    III.PROBLEM FORMULATION

    We can attempt to determine the migration decisionωrand migration policyωr,nto minimize the long-term weighted average cost for completing all request tasks in the proposed MEC framework during the total time periodT.First,we can compute the total latency and the total energy consumption for the request tasksre,cgenerated by the MWEein the time slott,which are derived as

    whereI(·)is the indicate function meetingI(·)=1 if the condition is true;otherwise,I(·)=0.

    To this end,the weighted average cost composed of the latency cost and the energy consumption cost for each request taskre,cof the MWEein time slottcan be derived by

    whereη1,η2are the weight parameters for the latency cost and the energy consumption cost,respectively,meetingη1+η2=1.

    Accordingly,the goal of the paper is to seek the optimal migrating and offloading decision-making policy to obtain the minimum long-term weighted average cost for all request tasks of the MWE during the whole periodT.Therefore,minimizing the long-term weighted average cost problem is defnied as follows:

    where C1 stands for the migration decision-making limitation condition,which is the binary variable.In terms of C2,request tasks can be offloaded to the cloud server when all edge servers do not host the service caching required by tasks,ωr,n(t) is set as 0.Request tasks can be migrated to the edge server who hosts service caching required by tasks.The range ofωr,n(t)is[1,2,...,N].Accordingly,one edge server regardless of the MEC server or UAV can only host limited server cachings whose number is less than that of the cloud server on limitation C3.C4 indicates that in each time slot,an MWE can only generate one request task at most;that is,there is a certain probabilityPr(t) of generating a request task.According to C5,all request tasks must be completed under a predefnied deadline.The number of allocated computation densities(CPU cycles)for executing the request task is less than the maximum capacity (achievable CPU cycles)of the corresponding edge server in C6.

    IV.MADRL-BASED COMPUTATIONAL OFFLOADING AND TASK MIGRATION SCHEME

    Lemma 1(Optimization problemP1in (18) is NP-hard under constraints C1-C6).

    Proof.We can see that problemP1in(18)is a mixedinteger programming that is diffciult to tackle which involves the select variableωr,nand 0-1 integer variableωr.Furthermore,we assumeωr,nis known,the residual sub-problem is a quadratic problem with 0-1 integer under constraints C1,C3-C6,which is a nondefniite matrix.0-1 programming problem is a typical and widely recognized NP-hard problem[34,35].It can be seen that subproblem is a NP-hard problem without considering the optimization variableωr,nunder constraints C1,C3-C6.

    Therefore,P1is a NP-Hard vexed question under constraints C1-C6.

    This problem commonly reformulated by specifci relaxation approach and then solved by powerful convex optimization techniques.However,P1is considered a multi-MWE cooperation and competition problem.To addressP1,we propose a MADRL-based computation offloading and task migration scheme.

    4.1 MDP Model

    First,we transit the computational offloading and task migration problem to the MDP,and each MWE is considered a learning agent.

    4.1.1 State Space

    In terms of the request tasksre,cfrom the MWEe,the state in time slottincludes the arrival probability of the request tasksre,c,the size of the request tasksre,c,the priority of the request tasksre,c,the length of buffer queue on edge servern,the channel state between MWEeand edge servern,the service caching hosted by edge servern,and the changing rate of capacity of edge serverλc.

    4.1.2 Action Space

    The actionae(t)∈Ais composed of the migrating decision and migrating or offloading selection,which can be written by,

    The agent chooses actionae(t)for each request taskre,cfrom MWEeand transits to the next state.

    4.1.3 Reward Function

    From(18),we consider the optimization problemP1and the limitation conditions C1 ∽C6.We can see thatP1is the global optimization scheme for request tasks of all MWEs.However,we defnie the immediate rewardre(t) as the reward indicate functionD(-aΓe-b)for the request taskre,cof MWEe,wherea,bare the constant coeffciients for the reward function.

    Γis the long-term weighted average cost for request taskre,cof MWEeduring the whole periodT.Accordingly,the maximum rewardre(t) can be considered the minimum weighted average cost for executing the request tasksre,cfrom the MWEe.

    4.2 Minimizing Cost Scheme Based on MADDPG

    The trouble faced by traditional single agent DRL algorithms such as DQN and DDPG is that since each agent can constantly learn and train to improve its optimal policy,from the perspective of each agent,the environment is dynamic and time-varying,and traditional DRL has diffciulty meeting convergence conditions.

    MADDPG is essentially a variant of the DDPG algorithm that trains the critic network that requires global information and an actor network that requires local information for each agent and allows each agent to defnie its own reward function,so it can be used for cooperative tasks or competing tasks[36].MADDPG algorithm has three advantages: 1) Centralized training,but distributed execution;2)Improved experience replay of samples;3) Use a policy ensemble to optimize performance.

    Therefore,using the advantage of centralized training and decentralized execution for MADDPG,we propose a MADDPG-based computation offloading and task migrating decision-making scheme(MBOMS) for minimizing the long-term average weighted cost in our proposed multi-MWEs MEC framework environment.

    We defnie the policy parameters and the policy of the MWE agent asυ=[υ1,υ2,...,υe] andπ=[π1,π2,...,πE],respectively.The expected cumulative rewards of agentecan be derived by

    Agentecan observe the joint state from environment composed of the states and the actions of all MWE agentsx=(o1,o2,...,oE),namely,the global observation isx,and the joint action is defnied asa=(a1,a2,...,aE).Because the selection of action follows the deterministic policy,the deterministic policy gradient of agenteis

    whereoeis the observation of the agente.is the centralized state-action function of agente.Since each agent learns its own state-action functionindependently,the reward function of each agent is different.Merbstands for the experience replay buffer of all agents,and each sample inMerbis defnied as the tuple〈x,x′,a1,a2,...,an,r1,r2,...,rn〉.

    In terms of the value-based critic,inspired by the time difference(TD)and target network of DDQN,the loss is calculated by the mean squared error(MSE)as follows:

    For the critic,the gradient can be updated by

    In terms of the T-network’s parameters,after certain time period,MADDPG can copy parameters from Enetwork to T-network for each agent of actor and critic as follows:

    As shown in Figure 2,the proposed MADDPGbased framework includesEagents.Each agent can interact with the proposed MEC environment,which is similar to the agent of DDPG when execution.However,when training the neural networks of MBOMS,centralized learning methods are utilized to train all critic networks and actor networks.For instance,during the centralized training process,agentecan observe the joint state from environment composed of the states and the actions of all MWE agentsx=(o1,o2,...,oE).During the time slott,the observed statexis input into the actor network of agente;then,the output of the actor network is the local optimal actionaeby the local policy.To this end,the immediate rewardrecan be obtained from the environment and transits to the next observationx′based on the joint actiona.The minibatch transition samples can be stored in the experience replay buffer.Furthermore,the chosen local actiona′e=υe(oe),the global observationx,and the joint actiona={a1,a2,...,aE}are input to the critic network of the agente.In the critic network,each agent can update the state-action functionQυeby training separately.The actions of the other agents can be obtained by interacting with the stationary environment due to offline processing.We can see that the centralized training of the proposed MBOMS algorithm can be convergent and address the unstable environment caused by the different actions of decentralized multiagent algorithms.When executed,each agent can only observe the local state from the actor network and execute the local action with no information of other agents.

    Figure 2.Structure of the proposed MBOMS.

    As illustrated in Figure 2,MBOMS can be described in Algorithm 1.First,the parameters used in MBOMS are initialized.Episodes are defnied to better learn the optimal offloading and migrating policy,and one episode includes the whole time periodT.

    4.3 Complexity Analysis

    For MADDPG-based task offloading service migrating scheme of proposed MBOMS,in terms of each WMEe ∈Ein each group,the neural network for MADDPG is composed ofLk+1 layers: an input layer,Lk-1 fully connected layers,and an output layer.We assume thatMis the number of training samples.Pedenotes the number of Epoch for computing the loss function.nIrepresents the dimension of input layer which is related to the size of state space.From (??),the size of state space iswhere |N| is the size of MEC server,is the size of service cachings hosted by MEC servern.nlis the number of neurons in layern,wheren ∈N,n ≥2.nlis the number of neurons in layern,wheren ∈ N,n ≥2.nOrepresents the dimension of output layer which is related to the size of action space.Therefore,the computation complexity of the proposed MBOMS isO(E×M×Pe×nI×(Lk-1)nl×nO).In addition,we ignored the complexity of all activation functionO(nl),which is only related tonland far less than the complexity of input layer,hidden layer,and output layer for MBOMS.

    V.NUMERICAL SIMULATION RESULTS

    In this section,we describe in detail the convergence and effectiveness of the proposed MADDPG-based offloading and migrating algorithm in a multi-MWE MEC framework via extensive simulations.

    5.1 Simulation Setup

    The main simulation parameters are listed in Table 2.In addition,for priorities 1,2,and 3 of request tasks,the arrival probability is set as [1/3,1/3,1/3].The service cachings hosted by each edge server are round robin varying in 100 episodes;that is,during one epoch (overall episodes),service cachings hosted by each edge server will change 10 times.The channel transition probability matrixAn,cis set as the same probability.The channel state vector is set asSchannel=sc* 28* 109,wheresc=[1,2,3,4,5,6,7,8,9].The size of request tasks is set asU~[1,5] Mb.The powers for allocation,transmission for MWEe,and edge servernare all set as 30 dBm.κnis set as 2*10-28.The coeffciients for indicator functionDare set asa=1,b=0.

    Table 2.Simulation parameter list of the proposed algorithm.

    The structures of the E-network and T-network for the actor network and critic network are two layers of fully connected networks.

    5.2 Performance Comparison

    Five benchmarks are introduced to evaluate and verify the performance of the proposed MADDPG-based offloading and migrating algorithm,which are defnied as:1)Decentralized DDPG-based scheme(DDBS);2)MADDPG-based scheme without migration(MBSM);3) Centralized DDPG-based scheme (CDBS);4) Decentralized DDQN-based scheme (DDQN);5) Centralized scheme based on Q-Learning(Q-Learning).

    5.3 Simulation Results

    The goal of our work is to minimize the weighted average cost (maximize the expected cumulative reward)by fniding the optimal offloading and migrating decision-making policy.First,we evaluate the convergence of our proposed MBOMS algorithm by the average expected cumulative reward during overall time periodTper episode,which is defnied as

    Figure 3 reveals the convergence of MBOMS with different learning rates of criticαcin terms of the average expected cumulative reward per episodeRAPE.Whenαc=0.1,MBOMS works poorly and cannot be convergent because the change in the learning step is too great.If the learning rate is set asαc=0.01,MBOMS can converge well according to the green curve.However,the learning curve fluctuates too much,especially after 800 episodes.Whenαc=0.001,the convergence and performance of MBOMS is better than that when the learning rate is set to other values.

    Figure 3.Convergence of MBOMS with different learning rates of critic.

    Figure 4 shows the performance of MBOMS in terms of the average expected cumulative reward per episodeRAPEwith different arrival probabilitiesProf request taskre,c.The average expected cumulative reward per episodeRAPEis signifciantly decreased with an increase inPr.However,the convergence of MBOMS can be guaranteed with different confgiurations for probabilityPr,which proves the robustness of our proposed MBOMS.The greatPrmakes the number of request tasks generated by MWE dramatically grow,which leads to the massive penalty during the overall time periodTin each episode.

    Figure 4.Convergence of MBOMS with different arrival probabilities of request tasks.

    Figure 5 illustrates the performance and convergence of six algorithms considering the average expected cumulative reward per episodeRAPE.We can see that the six algorithms can be convergent after several episodes,but the performance is dramatically different.However,the performance of MBOMS is better than that of DDBS,MBSM,and CDBS.Thanks to the decentralized learning pattern,DDBS can gain a better performance,but the average expected cumulative rewardRAPEsignifciantly fluctuates and becomes unstable compared with our proposed MBOMS.Due to no migration for all request tasks,RAPEof MBSM is less than that of MBOMS and DDBS,although the migration cost of MBOMS is great.The performance of CDBS is worst because the centralized training and execution,especially the massive number of requests,works from MWEs.

    Figure 5.Convergence of different algorithms.

    To evaluate and verify the robust performance of MBOMS,we illustrate the different capacities of edge servers in terms of different priorities of request tasks.Figure 6(a) assumes that the capacity of edge servers(MEC servers and UAVs) is 40.5 GHz.Beneftiting from task migration among the edge server,the unfniished ratio of MBOMS is less than that of MBSM regardless of the priority of request tasks in Figure 6(b).When capacity of all edge servers is set as 30.5 GHz,the number of unfniished request tasks increases slightly.However,the performance of MBOMS is better than that of MBSM considering the unfniished tasks ratio.As shown in Figure 6(c),when the capacity of edge servers (MEC servers and UAVs) is normal but the number of MWEs is set as 30,the number of dropped request tasks will violently increase due to insuffciient CPU cycles for massive tasks of MWEs.However,in terms of all request tasks with different priorities,the unfniished ratio of MBSM is still less than that of MBSM,and task migration is more important for completion under the most stringent deadline.

    Figure 6.Average unfinished ratio of request tasks with different priorities.

    Furthermore,we will compare with fvie other algorithms (DDBS,MBSM,CDBS,DDQN,and QLearning) to verify and confrim the advantages and effectiveness of our proposed MBOMS algorithm.We defnie the mean of expected cumulative rewards of all agents during overall episodes to simulate the negative weighted average cost of the MEC system,which is written as

    Figure 7 depicts the mean of expected cumulative rewardsRMOEof the six algorithms with different numbers of MWEsE.WhenE ≤20,the bandwidth and computational resources are suffciient to execute the offloaded request tasks from these MWEs,and theRMOEof the six algorithms does not change signif-i cantly.However,whenE >20,because the massive MWEs will generate more request tasks,which increases the burden of edge servers and leads to the massive penalty for unfniished tasks,RMOEdecreases dramatically.

    Figure 7.Comparison results with different numbers of MWEs.

    Figure 8 illustrates the relationship betweenRMOEand the maximum capacity of the MEC server.According to the maximum capacity of the MEC server shown in axisx,the value of UAV corresponding to the edge server is set as [20.5,25.5,30.5,35.5,40.5]GHz.We can see from the curves in Figure 8 that as the capacity of the edge server increases,RMOEgrows gradually due to the suffciient computation ability of the edge server.In terms ofRMOE,MBOMS outperforms DDBS,MBSM,CDBS,DDQN,and QLearning.When the capacity of MEC server is less than 35.5 GHz because of the competition of computational resources from the MEC server among offloaded request tasks,the performance of DDBS,DDQN are signifciantly worse than those of MBOMS and MBSM.However,the capacity of the MEC server grows gradually,and the performance of DDBS,DDQN are close to those of MBOMS and MBSM.Nevertheless,MBOMS has the best performance.QLearning has the worst performance due to its lack of learning and great Q-Table.

    Figure 8.Comparison results with different maximum capacities of the MEC server.

    The channel vector is an important factor that can influence the channel gain and uplink available transmission rate between the MWE and edge server.Figure 9 shows the effect of channel vectorSchannelbetween MWEeand edge servernfor the mean of expected cumulative rewardsRMOEof the six algorithms.The greater the channel vectorSchannelis,the greater the uplink available transmission rateWhen the index of channel vectorx <8,the lower transmission rate causes greater transmission latency and wait latency in the task buffer queue.Therefore,there are more unfniished request tasks under the deadline,which results in worse performance of the six algorithms.AlthoughRMOEincreases as the index of channel vectorx >8,our proposed MBOMS achieves the best performance because of the cooperative training pattern.

    Figure 9.Comparison results with different channel vectors.

    VI.CONCLUSION

    In this paper,we design a three-tier architecture MEC system assisted with service caching and propose an MADDPG-based computation offloading and task migrating decision-making scheme,named MBOMS.Based on MBOMS,the long-term average weighted cost of the overall MEC system can be minimized by training the neural networks.MBOMS is a robust centralized training and decentralized executing algorithm that considers the randomness of the request tasks generated by the MWE,the time-varying channel between the MWE and the edge server,the changing capacity of the edge server,and the different priorities of request tasks.In the simulation results,the convergence and effectiveness of MBOMS are systematically and comprehensively evaluated and verifeid.

    ACKNOWLEDGEMENT

    This work was supported by Jilin Provincial Science and Technology Department Natural Science Foundation of China (20210101415JC) and Jilin Provincial Science and Technology Department Free exploration research project of China(YDZJ202201ZYTS642).

    国产欧美日韩精品亚洲av| 精品卡一卡二卡四卡免费| 可以免费在线观看a视频的电影网站| bbb黄色大片| 免费不卡黄色视频| 久久午夜综合久久蜜桃| xxxhd国产人妻xxx| 免费不卡黄色视频| 国产av一区二区精品久久| 91老司机精品| 人人妻人人添人人爽欧美一区卜| 丁香六月欧美| 丁香六月欧美| 99久久99久久久精品蜜桃| 高清欧美精品videossex| 精品福利观看| 黑人猛操日本美女一级片| 两个人看的免费小视频| 国产亚洲欧美精品永久| 美女扒开内裤让男人捅视频| 精品第一国产精品| 99九九在线精品视频| 亚洲一区中文字幕在线| 12—13女人毛片做爰片一| 少妇精品久久久久久久| 亚洲欧美精品自产自拍| 美女扒开内裤让男人捅视频| 秋霞在线观看毛片| www.999成人在线观看| 成人免费观看视频高清| 成人国产av品久久久| 亚洲精品av麻豆狂野| 动漫黄色视频在线观看| 亚洲精品粉嫩美女一区| 国产欧美日韩综合在线一区二区| netflix在线观看网站| www.自偷自拍.com| 日本a在线网址| 一本色道久久久久久精品综合| 99九九在线精品视频| av线在线观看网站| 99国产综合亚洲精品| 老司机午夜十八禁免费视频| 国产免费一区二区三区四区乱码| 日本欧美视频一区| 久久精品成人免费网站| 日本av手机在线免费观看| 久久久久久免费高清国产稀缺| 欧美老熟妇乱子伦牲交| 国产黄频视频在线观看| 免费一级毛片在线播放高清视频 | 国产在线视频一区二区| 久9热在线精品视频| 欧美久久黑人一区二区| 纯流量卡能插随身wifi吗| 日韩,欧美,国产一区二区三区| 制服诱惑二区| 国产免费一区二区三区四区乱码| 精品一区二区三卡| 最近最新中文字幕大全免费视频| 亚洲av成人一区二区三| 久久精品国产a三级三级三级| 国产亚洲欧美在线一区二区| 首页视频小说图片口味搜索| 欧美精品亚洲一区二区| 男女床上黄色一级片免费看| 极品人妻少妇av视频| 亚洲国产欧美网| 99国产极品粉嫩在线观看| 国产亚洲欧美在线一区二区| 色婷婷av一区二区三区视频| 男人爽女人下面视频在线观看| 久久精品熟女亚洲av麻豆精品| 欧美精品一区二区大全| 亚洲av电影在线观看一区二区三区| 黑人巨大精品欧美一区二区mp4| kizo精华| 亚洲一卡2卡3卡4卡5卡精品中文| 男男h啪啪无遮挡| 丝瓜视频免费看黄片| 精品久久久久久久毛片微露脸 | 婷婷成人精品国产| 国产精品自产拍在线观看55亚洲 | 最黄视频免费看| 中文字幕av电影在线播放| 好男人电影高清在线观看| 午夜福利在线免费观看网站| 各种免费的搞黄视频| 免费日韩欧美在线观看| 男女无遮挡免费网站观看| 久久久久久久久久久久大奶| 中国国产av一级| 韩国高清视频一区二区三区| 久久女婷五月综合色啪小说| 老司机福利观看| 成年人午夜在线观看视频| 久久亚洲精品不卡| 久久久国产成人免费| 国产成人a∨麻豆精品| 精品人妻一区二区三区麻豆| 母亲3免费完整高清在线观看| 国产亚洲午夜精品一区二区久久| 少妇粗大呻吟视频| 国内毛片毛片毛片毛片毛片| 欧美少妇被猛烈插入视频| 秋霞在线观看毛片| 午夜福利在线免费观看网站| 午夜成年电影在线免费观看| 国产欧美日韩一区二区三区在线| 男女之事视频高清在线观看| av天堂久久9| 亚洲国产欧美日韩在线播放| 亚洲精品中文字幕一二三四区 | 久久人人爽人人片av| 精品乱码久久久久久99久播| 久久精品久久久久久噜噜老黄| 久久精品国产综合久久久| tube8黄色片| 久久久久久久久久久久大奶| 啦啦啦视频在线资源免费观看| 极品少妇高潮喷水抽搐| 视频在线观看一区二区三区| 脱女人内裤的视频| 91成人精品电影| 高清黄色对白视频在线免费看| 日韩一卡2卡3卡4卡2021年| av网站在线播放免费| 国产成人精品无人区| 色精品久久人妻99蜜桃| 国产精品偷伦视频观看了| 免费不卡黄色视频| 精品高清国产在线一区| 国产高清视频在线播放一区 | 精品一区二区三区四区五区乱码| 91精品国产国语对白视频| 亚洲精品国产av蜜桃| 一二三四社区在线视频社区8| 国产色视频综合| 国产高清国产精品国产三级| 久久久久国产精品人妻一区二区| av欧美777| 电影成人av| 各种免费的搞黄视频| 9191精品国产免费久久| 热re99久久精品国产66热6| 亚洲av美国av| 国产精品成人在线| 国产亚洲午夜精品一区二区久久| 老司机在亚洲福利影院| 亚洲精品一区蜜桃| av欧美777| 亚洲性夜色夜夜综合| 视频区图区小说| 50天的宝宝边吃奶边哭怎么回事| 秋霞在线观看毛片| 日韩视频在线欧美| 精品免费久久久久久久清纯 | 成年动漫av网址| 免费观看av网站的网址| av片东京热男人的天堂| 王馨瑶露胸无遮挡在线观看| 久久久久久久国产电影| 男女床上黄色一级片免费看| 大香蕉久久成人网| av天堂久久9| 亚洲精品第二区| 美女中出高潮动态图| 久久国产精品大桥未久av| 久久99一区二区三区| 18禁裸乳无遮挡动漫免费视频| 日日爽夜夜爽网站| 亚洲熟女毛片儿| 久久狼人影院| 大陆偷拍与自拍| 99久久国产精品久久久| 一本久久精品| 国产淫语在线视频| 国产成人影院久久av| 久久久久精品国产欧美久久久 | 日韩制服丝袜自拍偷拍| 国产精品九九99| 国产极品粉嫩免费观看在线| 亚洲色图综合在线观看| 妹子高潮喷水视频| 久久亚洲国产成人精品v| 色精品久久人妻99蜜桃| 国产男女内射视频| 国产老妇伦熟女老妇高清| 日韩熟女老妇一区二区性免费视频| 欧美成狂野欧美在线观看| av天堂久久9| 亚洲欧美激情在线| tube8黄色片| 国产精品一区二区在线观看99| 两性夫妻黄色片| 法律面前人人平等表现在哪些方面 | 亚洲欧美日韩另类电影网站| av欧美777| 国产成人精品久久二区二区91| 中文字幕色久视频| 搡老熟女国产l中国老女人| 伊人久久大香线蕉亚洲五| 十八禁网站网址无遮挡| 久久九九热精品免费| 国产一区二区 视频在线| 国产精品一二三区在线看| 亚洲成人国产一区在线观看| 亚洲精品乱久久久久久| 老熟妇仑乱视频hdxx| 狠狠狠狠99中文字幕| 亚洲精品久久午夜乱码| 欧美午夜高清在线| 日本撒尿小便嘘嘘汇集6| 亚洲精品第二区| 国产精品自产拍在线观看55亚洲 | 777米奇影视久久| 999精品在线视频| 天堂中文最新版在线下载| 我的亚洲天堂| 丝袜喷水一区| 18禁黄网站禁片午夜丰满| 老司机在亚洲福利影院| 午夜福利免费观看在线| 精品一区二区三区av网在线观看 | 欧美黑人欧美精品刺激| 大码成人一级视频| 亚洲精品一区蜜桃| 亚洲五月婷婷丁香| 欧美精品亚洲一区二区| 国产免费福利视频在线观看| 亚洲av成人一区二区三| 19禁男女啪啪无遮挡网站| 十分钟在线观看高清视频www| 午夜免费鲁丝| 99久久精品国产亚洲精品| 一级毛片精品| 久久人人97超碰香蕉20202| 精品国产一区二区三区久久久樱花| 精品熟女少妇八av免费久了| 制服人妻中文乱码| 国产精品1区2区在线观看. | 久久九九热精品免费| 国产日韩一区二区三区精品不卡| 午夜两性在线视频| 欧美黄色片欧美黄色片| 久久九九热精品免费| 午夜免费鲁丝| 婷婷成人精品国产| 精品少妇黑人巨大在线播放| 制服诱惑二区| 日日摸夜夜添夜夜添小说| 亚洲专区中文字幕在线| 色94色欧美一区二区| 最新的欧美精品一区二区| 黄色视频,在线免费观看| 如日韩欧美国产精品一区二区三区| 久久国产精品人妻蜜桃| 久久人人爽人人片av| 人人妻人人爽人人添夜夜欢视频| 国产色视频综合| 久久青草综合色| 韩国精品一区二区三区| 国产av精品麻豆| 狂野欧美激情性xxxx| 亚洲国产精品成人久久小说| 欧美变态另类bdsm刘玥| 久久精品亚洲av国产电影网| 国产一区二区 视频在线| 亚洲精品国产区一区二| videosex国产| 99热网站在线观看| 亚洲自偷自拍图片 自拍| 国产精品麻豆人妻色哟哟久久| 后天国语完整版免费观看| 曰老女人黄片| 大型av网站在线播放| 午夜日韩欧美国产| 欧美激情 高清一区二区三区| 色综合欧美亚洲国产小说| 国产精品影院久久| 一进一出抽搐动态| 12—13女人毛片做爰片一| 亚洲一区二区三区欧美精品| 国产精品99久久99久久久不卡| 欧美激情高清一区二区三区| 大片免费播放器 马上看| 美女高潮喷水抽搐中文字幕| 亚洲av电影在线观看一区二区三区| 美女午夜性视频免费| 菩萨蛮人人尽说江南好唐韦庄| 免费在线观看影片大全网站| 日韩大码丰满熟妇| 精品少妇一区二区三区视频日本电影| 久久99一区二区三区| 欧美日韩国产mv在线观看视频| 久久久久精品人妻al黑| 制服人妻中文乱码| 欧美黑人精品巨大| √禁漫天堂资源中文www| 国产一区有黄有色的免费视频| 日本wwww免费看| 亚洲精品国产区一区二| 如日韩欧美国产精品一区二区三区| 一二三四在线观看免费中文在| 亚洲欧美一区二区三区黑人| 午夜91福利影院| av有码第一页| 91精品伊人久久大香线蕉| 精品少妇黑人巨大在线播放| 日本av手机在线免费观看| 他把我摸到了高潮在线观看 | 中文字幕av电影在线播放| 人人妻人人爽人人添夜夜欢视频| 黑人欧美特级aaaaaa片| 欧美老熟妇乱子伦牲交| 精品第一国产精品| 99精品欧美一区二区三区四区| 一级片'在线观看视频| 国产野战对白在线观看| e午夜精品久久久久久久| 午夜精品国产一区二区电影| 99精品久久久久人妻精品| 热re99久久精品国产66热6| 亚洲av国产av综合av卡| 亚洲精品成人av观看孕妇| 久久久久久久久久久久大奶| 国产精品一区二区精品视频观看| 一级,二级,三级黄色视频| 美女国产高潮福利片在线看| 欧美在线一区亚洲| 日韩 亚洲 欧美在线| 日本a在线网址| 午夜福利影视在线免费观看| 精品国产乱码久久久久久男人| 亚洲精品av麻豆狂野| 手机成人av网站| 久久国产精品大桥未久av| 国产免费视频播放在线视频| 视频区欧美日本亚洲| 一区二区三区四区激情视频| 国产精品九九99| 亚洲 欧美一区二区三区| 亚洲男人天堂网一区| 91大片在线观看| 欧美午夜高清在线| 久久人人爽av亚洲精品天堂| 亚洲久久久国产精品| 丰满迷人的少妇在线观看| 蜜桃在线观看..| 精品一区在线观看国产| 精品乱码久久久久久99久播| 精品国产乱子伦一区二区三区 | 91九色精品人成在线观看| 国产免费现黄频在线看| 91九色精品人成在线观看| 国产精品自产拍在线观看55亚洲 | 日日夜夜操网爽| 日韩,欧美,国产一区二区三区| 精品国产乱子伦一区二区三区 | 又黄又粗又硬又大视频| 亚洲av日韩在线播放| www日本在线高清视频| 久久精品国产亚洲av高清一级| 又黄又粗又硬又大视频| 国产成人精品久久二区二区91| 在线 av 中文字幕| 老鸭窝网址在线观看| 国产精品一区二区精品视频观看| 十八禁人妻一区二区| 国产日韩欧美视频二区| 国产av又大| 欧美日韩黄片免| 亚洲天堂av无毛| 美女视频免费永久观看网站| bbb黄色大片| 成人国产一区最新在线观看| 亚洲av男天堂| 日韩视频在线欧美| 久久久久精品国产欧美久久久 | 高清av免费在线| 精品一品国产午夜福利视频| 亚洲精品国产色婷婷电影| 另类亚洲欧美激情| 国产一区二区 视频在线| 高清视频免费观看一区二区| 丁香六月欧美| 最近中文字幕2019免费版| 国产亚洲精品一区二区www | 欧美日韩亚洲综合一区二区三区_| av电影中文网址| 国产伦人伦偷精品视频| 下体分泌物呈黄色| a在线观看视频网站| 99热全是精品| 国产欧美日韩一区二区精品| 天天躁夜夜躁狠狠躁躁| 久久九九热精品免费| 亚洲国产精品成人久久小说| 18禁观看日本| 亚洲欧美色中文字幕在线| 久久精品成人免费网站| 国产成人精品久久二区二区免费| 日韩电影二区| 操出白浆在线播放| 少妇 在线观看| 欧美激情高清一区二区三区| 亚洲精品久久午夜乱码| 精品国产一区二区三区四区第35| 欧美日韩福利视频一区二区| a 毛片基地| 最近中文字幕2019免费版| 天堂8中文在线网| 国产成人系列免费观看| 中文字幕人妻熟女乱码| 国产精品一区二区精品视频观看| 久久国产精品大桥未久av| 大片电影免费在线观看免费| 亚洲伊人色综图| 亚洲熟女精品中文字幕| 国产精品久久久久久精品电影小说| 最近中文字幕2019免费版| 高清欧美精品videossex| cao死你这个sao货| 亚洲精品国产av成人精品| 国产精品.久久久| 亚洲精品乱久久久久久| 国产亚洲欧美精品永久| 精品国产超薄肉色丝袜足j| 99精国产麻豆久久婷婷| 成人18禁高潮啪啪吃奶动态图| 精品国产超薄肉色丝袜足j| 亚洲欧美激情在线| 天天添夜夜摸| 中文字幕最新亚洲高清| 日本91视频免费播放| 后天国语完整版免费观看| 超碰成人久久| 久久久水蜜桃国产精品网| 亚洲国产中文字幕在线视频| 亚洲欧美色中文字幕在线| 国产欧美日韩一区二区精品| 欧美精品亚洲一区二区| 国产av精品麻豆| 日本av手机在线免费观看| 午夜免费观看性视频| 女人精品久久久久毛片| 悠悠久久av| 桃红色精品国产亚洲av| 国产老妇伦熟女老妇高清| 亚洲伊人色综图| 久久九九热精品免费| 男人舔女人的私密视频| 日韩中文字幕视频在线看片| 亚洲精品一二三| 啦啦啦中文免费视频观看日本| 久久久久精品人妻al黑| 久久国产精品人妻蜜桃| 国产主播在线观看一区二区| 少妇人妻久久综合中文| 欧美日韩黄片免| 亚洲美女黄色视频免费看| 不卡一级毛片| 久久九九热精品免费| 国产无遮挡羞羞视频在线观看| 如日韩欧美国产精品一区二区三区| 久久久久久免费高清国产稀缺| 嫁个100分男人电影在线观看| 亚洲欧美激情在线| 免费高清在线观看视频在线观看| 亚洲五月色婷婷综合| 亚洲一码二码三码区别大吗| 精品高清国产在线一区| 日韩,欧美,国产一区二区三区| 精品高清国产在线一区| 天天操日日干夜夜撸| 热re99久久精品国产66热6| 国产精品秋霞免费鲁丝片| 欧美97在线视频| 国产精品一区二区免费欧美 | 91大片在线观看| 欧美激情极品国产一区二区三区| 中文字幕精品免费在线观看视频| 免费看十八禁软件| 成年av动漫网址| 亚洲第一欧美日韩一区二区三区 | 女性生殖器流出的白浆| 老司机亚洲免费影院| 亚洲激情五月婷婷啪啪| 久久久久久久国产电影| 亚洲国产欧美在线一区| 亚洲第一青青草原| 9191精品国产免费久久| 国产精品 国内视频| 欧美亚洲日本最大视频资源| 欧美黄色淫秽网站| 国产又色又爽无遮挡免| 在线观看免费高清a一片| 亚洲欧美精品自产自拍| 丰满饥渴人妻一区二区三| 视频区图区小说| 美女午夜性视频免费| 9色porny在线观看| 精品国产一区二区三区久久久樱花| 97人妻天天添夜夜摸| 麻豆国产av国片精品| 亚洲成人免费av在线播放| av一本久久久久| 19禁男女啪啪无遮挡网站| av一本久久久久| www.av在线官网国产| 色婷婷久久久亚洲欧美| 亚洲天堂av无毛| 在线观看一区二区三区激情| 人人澡人人妻人| 国内毛片毛片毛片毛片毛片| 建设人人有责人人尽责人人享有的| 精品久久久久久电影网| 国产欧美日韩一区二区三 | 精品少妇黑人巨大在线播放| 欧美 亚洲 国产 日韩一| 国产成人影院久久av| 国产在线一区二区三区精| 国产精品欧美亚洲77777| 国产一区二区激情短视频 | 国产欧美日韩一区二区精品| 国产日韩欧美在线精品| 精品少妇久久久久久888优播| 国产伦人伦偷精品视频| 蜜桃国产av成人99| 日韩中文字幕视频在线看片| 十八禁网站网址无遮挡| 国产老妇伦熟女老妇高清| 国产熟女午夜一区二区三区| 欧美在线黄色| av在线播放精品| 国产成人精品无人区| 少妇粗大呻吟视频| 亚洲精品国产一区二区精华液| 啦啦啦 在线观看视频| 色老头精品视频在线观看| 中文字幕人妻熟女乱码| 久久亚洲精品不卡| 丝袜美足系列| 免费高清在线观看视频在线观看| 91成人精品电影| 人人澡人人妻人| 51午夜福利影视在线观看| 亚洲精品国产区一区二| av网站在线播放免费| 国产在线免费精品| 一二三四在线观看免费中文在| 久久久精品国产亚洲av高清涩受| 亚洲视频免费观看视频| 日韩免费高清中文字幕av| 桃红色精品国产亚洲av| 亚洲欧美色中文字幕在线| 免费在线观看日本一区| 伊人亚洲综合成人网| 黑人巨大精品欧美一区二区蜜桃| 国产人伦9x9x在线观看| 免费高清在线观看日韩| 性色av一级| 中文精品一卡2卡3卡4更新| 国产欧美日韩精品亚洲av| 美女国产高潮福利片在线看| 国产成人一区二区三区免费视频网站| 日韩欧美一区二区三区在线观看 | 少妇的丰满在线观看| 十八禁网站免费在线| 男人爽女人下面视频在线观看| bbb黄色大片| kizo精华| 午夜日韩欧美国产| 高清黄色对白视频在线免费看| 亚洲精品成人av观看孕妇| 亚洲成人免费电影在线观看| 欧美日韩成人在线一区二区| 老汉色∧v一级毛片| 69精品国产乱码久久久| 丁香六月天网| 国产精品欧美亚洲77777| 高清视频免费观看一区二区| 啦啦啦中文免费视频观看日本| 国产亚洲欧美在线一区二区| 欧美国产精品va在线观看不卡| 超色免费av| av天堂在线播放| 欧美国产精品一级二级三级| 欧美黄色片欧美黄色片| 国产成人a∨麻豆精品| 天天操日日干夜夜撸| 亚洲欧美激情在线| 久9热在线精品视频| 午夜免费成人在线视频| 国内毛片毛片毛片毛片毛片| 2018国产大陆天天弄谢| 国产精品久久久人人做人人爽| 9色porny在线观看| 日韩熟女老妇一区二区性免费视频| netflix在线观看网站| 国产成人精品久久二区二区91| 日韩一区二区三区影片| 亚洲天堂av无毛| 涩涩av久久男人的天堂| 国产亚洲午夜精品一区二区久久| 亚洲成av片中文字幕在线观看| 亚洲av成人一区二区三| 纵有疾风起免费观看全集完整版| 久久久精品区二区三区| 亚洲国产av影院在线观看| 国产淫语在线视频| 国产成人av激情在线播放| 日韩视频一区二区在线观看| 老司机在亚洲福利影院| 久久香蕉激情|