• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep reinforcement learning-based optimization of lightweight task offloading for multi-user mobile edge computing

    2021-12-21 14:10:16ZHANGWenxianDUYongwen

    ZHANG Wenxian, DU Yongwen

    (School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou 730070, China)

    Abstract: To improve the quality of computation experience for mobile devices, mobile edge computing (MEC) is a promising paradigm by providing computing capabilities in close proximity within a sliced radio access network, which supports both traditional communication and MEC services. However, this kind of intensive computing problem is a high dimensional NP hard problem, and some machine learning methods do not have a good effect on solving this problem. In this paper, the Markov decision process model is established to find the excellent task offloading scheme, which maximizes the long-term utility performance, so as to make the best offloading decision according to the queue state, energy queue state and channel quality between mobile users and BS. In order to explore the curse of high dimension in state space, a candidate network is proposed based on edge computing optimize offloading (ECOO) algorithm with the application of deep deterministic policy gradient algorithm. Through simulation experiments, it is proved that the ECOO algorithm is superior to some deep reinforcement learning algorithms in terms of energy consumption and time delay. So the ECOO is good at dealing with high dimensional problems.

    Key words: multi-user mobile edge computing; task offloading; deep reinforcement learning

    0 Introduction

    Mobile devices nowadays are equipped with cameras, microphones and many other high quality sensors. Using these high-data rate sensory devices, it becomes possible for mobile devices to host perception related applications, such as face/gesture recognition, visual text translation and video image processing. While the popularization of intelligent devices and the emergence of 5G technology bring people comfort and convenience, they also cause many problems. For example, some mobile applications have very limited computing power when dealing with compute intensive tasks. As a result, to meet the quality of experience (QoE) of these mobile applications, the technology of mobile edge computing (MEC) has been proposed as a promising solution to bridge the gap between the limited resources on mobile devices and the ever-increasing demand of computation requested by mobile applications.

    According to Cisco’s report[1], mobile data traffic will grow sevenfold over the next five years, and reach 49 exabytes per month by 2021. Edge computing comes in handy at this point, aiming to reduce data transmission distances, which eliminates bandwidth and latency issues, improves application and service performance, power and reliability, and reduces operational costs ultimately.

    Task offloading solves the shortcomings of mobile devices in resource storage, computing performance and energy efficiency as a key technology of MEC. Task offloading technology not only reduces the pressure of core network, but reduces the time delay caused by transmission. It first uses the offload decision to determine which of computations to be processed locally and which to be uploaded to the edge server. The resource allocation is then calculated to determine where the final computing task will be unloaded.

    In a wireless fading environment, the time varying wireless channel condition largely impacts the optimal offloading decision of a wireless powered MEC system[2]. In a multi-user scenario, a major challenge is the joint optimization of individual computing mode (i.e., offloading or local computing) and wireless resource allocation (e.g., the transmission air time divided between wireless power transfer (WPT) and offloading). Such problems are generally formulated as mixed integer programming (MIP) problems due to the existence of binary offloading variables. To tackle the MIP problems, branch and bound algorithms[3]and dynamic programming[4]have been adopted, however, they require prohibitively high computational complexity, especially for large-scale MEC networks. To reduce the computational complexity, heuristic local search[5-6]and convex relaxation[7-8]methods are proposed. However, both of them require considerable number of iterations to reach a satisfying local optimum. Hence, they are not suitable for making real-time offloading decisions in fast fading channels, as the optimization problem needs to be re-solved once the channel fading has varied significantly.

    Conventional reinforcement learning (RL) algorithms cannot scale well as the number of agents increases, since the explosion of state space will make traditional tabular methods infeasible[9]. Nevertheless, by exploiting deep neural networks (DNNs) for function approximation, deep reinforcement learning (DRL) has been demonstrated to efficiently approximate Q-values of RL[10]. DRL combined with MEC enables mobile devices to learn optimal task offload decisions and energy allocation schemes based on task queue status, energy maximization long-term utility, queue status and channel quality.

    This paper considers how to determine more effectively whether computing tasks need to be offloaded to edge nodes. A multi-user MEC network is established, in which each user follows a binary offloading strategy with the goal of jointly optimizing user task offloading decisions according to time-varying wireless channels. Therefore, a computational offloading framework is proposed based on DRL to realize the offloading with the lowest time delay and energy consumption. Compared with existing algorithms based on DRL, the following novel contributions are obtained.

    1) Adopting an algorithm based on continuous action space on the premise of the existing decision making based on discrete action space to achieve a better power control of local execution and task offloading.

    2) Each mobile user with random task arrival and time-varying wireless channel independently learns dynamic offloading strategy in the multi-user MEC system to minimize power consumption and computing cost and reduce delay.

    3) Through experimental simulation, it is shown that edge computing optimize offloading (ECOO) has better simulation results than the learning performance of the traditional deep Q networks (DQN) and deep deterministic policy gradient (DDPG) algorithms in the decentralized strategy, and analyses the power delay trade-off of each user.

    1 Related work

    Artificial intelligence has brought great benefits to the society in the resource management of wireless network. DRL algorithm is proposed[11]to study the optimal caching and interference alignment in real time-varying wireless environment. Chen et al.[12]estimated which MEC could be used for task offloading on behalf of mobile users in the ultra-dense slice wireless access network with multiple base stations, and proposed a computational offloading algorithm strategy based on DQN to obtain the best choice of offloading project and maximize the long-term utility performance.

    There are many related works that jointly model the computing mode decision problem and resource allocation problem in MEC networks as the MIP problems. For instance, a coordinate descent (CD) method was proposed[5]that searched along one variable dimension at a time. A similar heuristic search method for multi-server MEC networks was studied[6], which iteratively adjusted binary offloading decisions. Another widely adopted heuristic was through convex relaxation, e.g., by relaxing integer variables to be continuous between 0 and 1[7]or by approximating the binary constraints with quadratic constraints[8]. Nonetheless, on one hand, the solution quality of the reduced-complexity heuristics is not guaranteed. On the other hand, both search-based and convex relaxation methods require considerable number of iterations to reach a satisfying local optimum and are inapplicable for fast fading channels.

    In recent years, research on multi-user computing offload in orthogonal frequency division multiple access (OFDMA) system in MEC environment[13]is concerned. Li et al.[14]proposed a partial data offloading algorithm on the basis of the research of multiuser computational shunt performed by orthogonal frequency division system to jointly optimize two subcarriers and power and apply them to the OFDMA MEC environment, and reduce latency on most mobile devices. Zhang et al.[15]studied the power and bandwidth allocated to OFDMA heterogeneous small cell networks to improve energy efficiency.

    In order to deal with the problems brought by random task arrival and time-varying wireless channels, the design of dynamic joint control strategy for radio and computing resources in MEC system becomes more challenging[16-19]. The partial computational shunt of multi-user is taken into account, and the resource allocation based on TDMA and OFDM is studied, so as to minimize the weighted sum of user energy consumption[16]. A multi-input/output system is designed, and the energy consumption problem of task offloading is solved through the joint optimization of the formed multi-input/output beam and the allocation of computational resources[17]. Green MEC system for energy acquisition is studied, in which delay cost is used to solve the problems of execution delay and task failure[18]. And the power delay trade-off in a multi-user scenario is discussed[19]. In the existing works, there have only strategies focusing on centralized DRL-based algorithms for optimal computation offloading in MEC systems, and the design of decentralized DRL-based algorithms for dynamic task offloading control of a multi-user MEC system still remains unknown.

    2 Problem model

    As shown in Fig.1, a multi-user MEC system has been established, consisting of a MEC server, a basic station (BS) and a group of mobile users, whereNis the number of users,N={1,2,…,N}. Each user in the system needs to complete computationally intensive tasks.

    Fig.1 Multi-user MEC system

    The MEC server near BS is deployed to improve the QoS of users through different user requirements considering the limited computing power of each mobile device. In addition, with the increase of the number of mobile users, multi-user MEC system makes it more convenient to solve the problem of offloading the decentralized tasks of each user, reduces the system overhead between users and MEC server, and improves the scalability of traditional MEC system.

    2.1 Network model

    A 5G macro cell or BS cell is designed as a MEC system, and linear detection algorithm is adopted to manage multiple mobile user uplink transmission (ZF). For each time slottbelongs tot, the received signal of BS can be written as

    (1)

    wherep0,n(t) is the transmission power of offloading task data bit of usern;sm(t) is the complex data symbol with unit variance, andn(t) is a Gaussian noise vector (AWGN) with zero mean and varianceσ2. In order to represent the correlation of slot time of each mobile user, the Gaussian Markov block fading autoregression model[20]is adopted.

    (2)

    whereρnis the normalized channel correlation coefficient between time slottandt-1, ande(t) is the error vector.

    (3)

    Therefore, the corresponding signal to interference plus noise ratio (SINR) is

    (4)

    It can be verified from Eq.(4) that the SINR of each user decreases as the number of users increases. More offloading power is needed to be allocated for each user to make the problem solved.

    2.2 Task model

    There are two types of task relationships: sequential tasks and concurrent tasks. For sequential tasks, the completion time can be reduced by offloading tasks to the edge cloud which has a larger processing capacity than the mobile device. For concurrent tasks, in addition to taking advantage of the greater processing power of the edge cloud, the parallelism of the mobile device and the edge cloud can also be exploited.

    1) Offloading sequential tasks

    To reduce the completion time of a sequential task-graph, tasks are needed to be offloaded as many to the cloud as possible to exploit the edge cloud’s greater processing capacity. However, offloading a task incurs a network data transmission which needs to be less than the time saved by offloading the task, in order to reduce completion time. In some situations, the network bandwidth can be so unfavorable that offloading tasks to the edge cloud will take even longer than running the task on the mobile device. It was proven that the optimal set of tasks to be offloaded in a sequential task-graph would always be a sequence of consecutive tasks in the task-graph[21]. Taking task-graph in Fig.2 as an example, the theorem in Ref.[21] states that if there are tasks that can be offloaded to the cloud to reduce the completion time, the tasks must be consecutive, that is, the tasks fromjtokin Fig.2.

    Fig.2 Offloading of sequential tasks

    2) Offloading concurrent tasks

    Optimizing the completion time of concurrent tasks is much more challenging. For a task-graphG(V,E), assume it begins with a fork task 0 and terminates with a merging taskn. The first fork task and the last merging task are usually local tasks, because the first task (which is called as the root task) needs to take inputs from the local device and the last task (which is called the termination task) needs to produce output back to the local device. The discussion of general task-graphs is left to the next subsection, where there are some sequential tasks before the root task, and some sequential tasks after the termination task.

    To offload concurrent tasks, parallelism is maximized between the mobile device and the cloud, which is equivalent to minimizing the completion time. Take Fig.3 as an example, where there arenconcurrent tasks between the root task 0 and the terminating taskn+1. If too many tasks are offloaded to the cloud, then the mobile device must wait for the edge cloud to finish completing its share of the tasks. If too few tasks are offloaded to the edge cloud, then the application would wait for the mobile device to complete its tasks, which prolongs the completion time. The best case result of an offloading decision is where the waiting time at the termination task, between tasks executed on the edge cloud and tasks executed on the mobile device, is as small as possible. To achieve this goal, when computing task offloading, the completion time of the local tasks must be as close as possible to the completion time of the offloaded tasks plus the data transmission delays.

    Fig.3 Concurrent tasks

    2.3 Computation model

    In this section, how each mobile user can meet the requirements of running their program by performing computing tasks locally or by offloading tasks will be discussed. Assuming all the applications are fine-grained[22],d1,m(t) means that the computation task is on the local mobile device, andd0,n(t) means that the computation task is offloaded to the edge server for execution. At the beginning of slott, the queue length of usern’s task buffer is

    Bn(t+1)=[Bn(t)-(d1,n(t)+d0,n(t))]++an(t),

    ?t∈Τ,

    (5)

    wherean(t) represents the number of task arrivals during time slott.

    1) Local computation:

    p1,n(t)∈[0,p1,n] is the power allocated to local execution. First of all, assuming that the number of CPU cycles required by usernto handle a task isLn, and the cycle number can be estimated by offline measurement[23]. Then the chip voltage is adjusted by using DVFS technology[24], and is the CPU frequency written on the effective switching capacitor in time slott. Therefore, the local processing int-slot can be derived as

    (6)

    (7)

    2) Edge computation

    Traditional MEC server is equipped with sufficient computing resources to solve the problems with edge computing, and all tasks offloaded to the MEC server through BS will be processed. Therefore the bit amount of offloading data of userncan be obtained according to Eq.(4) and expressed as

    d0,n=τ0Wlog2(1+γn(t)),

    (8)

    whereWis the system bandwidth.

    2.4 Energy consumption model

    The total energy consumption of all computing devices during a certain execution time, including smartphones, sensors and remote servers, is mainly composed of two parts: computing energy consumptionEcompand energy consumptionEofffor uninstalling mobile devices. Energy consumption model I can be calculated by

    (9)

    (10)

    wherealltaskiis the number of computing devices withitasks;Mjis the required CPU resources, andDRiis the total CPU resources.

    2.5 Cost model

    Due to the pay for the computing resources provided by the remote server, a new cost model is proposed based on the dynamic price model of the amount of remaining resources. The lower the amount of remaining resources is, the higher the price is. Under such a premise, users are more willing to choose the service node with a lower unit price as the offloading target, which helps reduce the use cost and improve the utilization rate of resources. The dynamic price model of residual resources in unit timetis

    (11)

    whereC1is the cost of the current equipment;U1is the interval time between the calculated expenses;R1is the univalence of the calculated resources;TMis the total computing resources of the current equipment, andL(t) is the computing resource ratio of the current equipment consumed in unit time. Since the computing resources of the local devices belong to the users themselves, no cost is needed to calculate. Therefore, the total cost of all remote devices is

    (12)

    3 Dynamic task offloading based on DRL

    RL makes the best decisions in a particular scenario through constant interaction with the environment. Based on continuous learning, RL can be more adaptable to the environment. While RL has many advantages, it lacks scalability and is limited to low dimensions.

    In order to solve the decision making difficulties encountered in reinforcement learning, DRL combines the perception ability of deep learning with the decision making ability of reinforcement learning, and solves the environmental problems of high dimensional state space and action space by relying on powerful functional approximation and the representation learning characteristics of deep neural network[25]. The overall optimization design of this paper is as shown in Fig.4.

    Fig.4 Dynamic offload system

    In the work, an improved DDPG algorithm is adopted[26]as shown in Fig.4, where each user independently learns the decentralized dynamic task offloading strategy after receiving the SINR and channel state information (CSI) sent by the base station, and provides computing tasks local execution and task offloading allocate power. In the DRL module, the neural network is improved: add a deep neural network and LSTM to the last layer of the actor network. Finally, join the candidate set network to ensure the optimal learning strategy. In particular, each user does not have prior knowledge of the MEC system, and therefore does not know the total number of users, that is, the value ofN. Then the ECOO algorithm is explained in detail.

    3.1 DRL framework

    State space: the MEC system needs to be observed comprehensively for the sake of the fully consideration of the characteristics of subtasks and server resources in the MEC system, including the channel vectors of all users and the queue length of task buffer. Whereas the real-world overhead of collecting such information on the BS and distributing them to users is enormous. It is used to reduce overhead and make the MEC system more scalable by letting each user perform operations in their own independent state.

    At the beginning of slott, the queue length of each user’s data bufferBn(t) will be updated according to Eq.(4), and the SINR finally received by BS will be passed to the user as feedback, and then channel vectorhn(t) of the upcoming uplink transmission will be estimated through channel reciprocity. Therefore, the state space is defined as

    Sn,t=[Bn(t),φn(t-1),hn(t)],

    (13)

    (14)

    For the purpose of ensuring that the offloading decision can be executed on the local mobile devices or remote servers, the offloading decision of the subtask needs to considerN+M+1 computing devices merely, including 1 cloud data center,Nlocal mobile devices andMedge servers.

    Action space: according to the current state of the system observed by each agentSn,t, the allocated power operationan,tthat selects local execution or task offloading for each time slottis

    an,t=[P1,n(t),P0,n(t)].

    (15)

    Different from other traditional DRL algorithms which select from several predefined discrete power levels, the improved DDPG algorithm can be applied to optimize power allocation from the continuous action space, which can significantly compensate for the high dimensional defects of the discrete action space.

    Reward function: the behavior of each agent is driven by reward. In order to learn the MEC model of energy perception dynamic calculation offloading strategy, the energy consumption is considered to be minimized within the acceptable buffer delay required to complete the task. According to Little theorem[27], the average queue length of the task buffer is proportional to the buffer delay, so the definition of the reward function that each agent receives after slottis

    rn,t=-ωn,1P1,n(t)-ωn,2P0,n(t)-ωn,3Bn(t).

    (16)

    Among them,ωn,1,ωn,2andωn,3are all nonnegative weighted factors. By assigning different values to them, the energy consumption and buffer delay during task offloading can be adjusted dynamically. The value function of the improved ECOO starting from the initial state to maximize usernunder strategyonis

    (17)

    Whenγ→1, the value function can be used to estimate the wireless range undiscounted return of each agent[28], and the average calculated cost is

    Cn(sn,t)=

    3.2 Optimization of candidate network

    The pseudo code of the ECOO algorithm proposed in this article is as follows.

    ECOO algorithm

    ECOO algorithm ensures the parameter difference between the current network and the target network through delayed update, so as to improve the stability of the training process. When the movement value in the training process is overestimated due to noise or error, the corresponding movement value will be overestimated inevitably in the later parameter updating process. Considering the results of multiple candidate networks comprehensively, action selection and action value evaluation are separated to ensure the optimal learning strategy[29].

    Fig.5 summarizes the components and features of DRL-based mobile edge caching, namely high-dimensional state space representation, neural network structure design and long-term reward maximization.As shown in Fig.5, the offloading agent observes the environment and obtains serval raw signals, such as user requests, context information and network conditions. These signals can be assembled into a high-dimensional state input and then fed to the deep neural network. The deep neural network needs to be designed into a specific structure like a convolutional neural network or a recurrent neural network, which is able to mine useful information and output the value function or the policy. According to the output, an action, which represents the computing offloaded at the next slot, can be selected. The resulting caching performance is then observed and passed back to the offloading agent as a reward. The offloading agent uses the reward to train and improve its deep neural network model with the aim of maximizing the expected accumulated discount reward.

    Fig.5 Deep reinforcement learning compute offloading process

    All computing tasks are assumed to have the same size, and the computing capacity of the edge node is assumed to be enough for computingCtasks. The edge node can serve each requesttdirectly if the requested task has been computing locally. Otherwise, the edge node requests this task from the original server and updates the local compute according to the computing policy. The aim is to find the optimal computing policy maximizing the offloaded traffic, that is, the number of computing answered by the edge node. The detailed functionalities of these networks are shown in Fig.6.

    Fig.6 Neural network architecture

    Upon receivingst, the computing agent needs to take an action at that corresponds to whether or not to compute the currently requested content in the task, and if yes, the agent determines which local servers will be compute task.

    In view of the gradual change of resources over time in MEC and the memory ability of LSTM network for long-term state, this paper proposes to combine LSTM and DDPG to deal with the time-dependent task offloading problem. The recurrent structure is used to integrate any long-term historical data to estimate the current state more accurately by replacing the last fully connected layer of the DDPG network with the LSTM layer.

    As shown in the Fig.7, it is assumed that candidate network setNet=(net1,net2,…,neti,…,netn) can store networks with a total number ofnandmnetworks with a network set ofNet, and update them after meeting a fixed number of iterationsC. Network settingsNet2has (n-m) net-works that are selected for updating by comparing the reward values. When the number of network settingsNet2is less than (n-m), the current network generated by each iteration is added toNet2as a candidate network. When the number of network settingsNet2equals (n-m), the current network and all network settingsNet2in the network will train the currently selected state-action pair; ifNet2is greater than (n-m), the candidate network with the minimum reward value is replaced by the current network, otherwise the training continues.

    Fig.7 ECOO algorithm training flowchart

    4 Simulation

    In this section, some numerical values of decentralized dynamic task offloading will be given in the MEC system, and the advantages and disadvantages of offloading decision are found out by comparing the cost, energy consumption and service delay. The implemented algorithms in large scale heterogeneous clusters include greedy Local execution first (GD-local), greedy computation first offloading (GD-offload), DQN-based dynamic offloading (DQN), DDPG-based dynamic offloading (DDPG) and ECOO.

    4.1 Parameter settings

    In the MEC system, the interval isτ0=1 ms. At the beginning of each iteration, each user’s channel vector is initialized tohn(0)~CN(0,h0(d0/dn)αIN), where path loss constantd0=1 m, path loss indexα=3, channel correlation indexρn=0.95, error vectore(t)~CN(0,h0(d0/d)αIN),fd,n=70 Hz. The system bandwidth is set as 1 MHz, the maximum transmitted power isP0,n=2 W, and noise power isσ2=10-9W. For local execution, assumeκ=10-27, CPU cycle required per bit isLn=500, and the maximum allowable CPU cycle frequency isFm=1.26 GHz. The maximum power required for local executionisP1,n=2 W.

    In DDPG algorithm, for each agent, actor network and critic network, there are two hidden four-layer fully connected neural networks using Relu activation function, and the number of neurons in the two hidden layers is 400 and 300, respectively. In order to implement ECOO algorithm, an experience replay buffer with a size of 10 000 is set up, which can return a randomly selected small batch of experiences when querying. In order to realize the ECOO algorithm, an experience replay buffer is set up with a size of 10 000, so that the randomly selected small-batch experience can be returned when querying, and the small-batch value can be set to 64, so as to realize the optimization of candidate network. At the same time, the adaptive moment estimation (Adam) method[30]is adopted and the learning rate is 0.001, respectively, and the soft update rate of the target network ist=100. In order to better explore the best offloading decision, set[31]θ=0.15 andσ=0.12[32]in Ornstein-Uhlenbeck process to provide relevant noise, and set the buffer size of experience playback as |Bn|=2.5×105.

    4.2 Multi-user simulation results

    In the MEC system, there are 5 mobile users, each of which is randomly located within 100 m distance from BS, and the task arrival rate isλn=n*1.0 Mbps. In the training stage, for different task arrival rates ranging from 1 Mbps-5 Mbps. The actor and critic network will be trained with the same network architecture and hyper-parameters. To compare the performance of different policies, testing results are averaged from 2 500 episodes, respectively.

    Fig.8 shows the training process of user dynamic computing offloading.

    Fig.8 Training process

    The reward value is the average of numerical simulation obtained in 10 times training sessions, where the task arrival rate is set asλ=3.0 Mbps. It can be observed from the experimental results that the average reward for iterations increases as the incremental number of interactions between the user agent and the MEC environment, which indicates that the ECOO algorithm can successfully learn efficient computing strategies without any prior know-ledge. In addition, the performance of strategies learned according to ECOO algorithm is always better than that of DDPG algorithm in the same scenario, manifesting that the strategies obtained based on ECOO can explore the action space more effectively than those obtained based on DDPG for continuous control problems.

    It can be observed from Fig.9 that the average reward will increase as the task arrival rate grows, which indicates that the computation cost is higher power consumption and a longer buffering delay.

    (a) Average reward

    The GD-local can achieve good results in latency, but the performance of cost and power. This is mainly because the GD-local algorithm prefers to offloads subtasks to the local device for execution. When resources of the local device are insufficient, subtasks are gradually offloaded to upper devices. Since some subtasks can be executed locally without network transmission, the GD-local algorithm has lower network latency and network usage. In addition, the GD-offload based on all computing tasks performed at the edge server is similar to the GD-Local algorithm, which consumes a lot of energy. The main reason is that the GD-offload algorithm tends to offload subtasks to edge server clusters, which consumes a lot of energy during the transfer process. At the same time, the performance of the edge server can meet the processing requirements of more subtasks, and improve the network usage of the entire cluster.

    The DQN algorithm, DDPG algorithm and ECOO algorithm all use the DRL to automatically generate the corresponding offloading strategy from the value iteration. As can be seen from the results in the Fig.9, as the task arrival rate increases, the improved DDPG algorithm outperforms the former two in terms of cost, power consumption and latency. It is because the ECOO algorithm comprehensively considers the historical parameters of the target network, and continuously updates the network parameters in real time, replacing the network with the minimum reward value, so as to keep the results always the best. At the same time as the lowest energy consumption, the buffer delay is always compromised.

    Testing results for the power-delay trade off is investigated by setting different values ofw1in Fig.10.

    Fig.10 Power-delay trade off

    It can be inferred from the curves that, there is a trade-off between the average power consumption and the average buffering delay. Specifically, with a largerw1, the power consumption will be decreased by sacrificing the delay performance, which indicates that in practicew1can be tuned to have a minimum power consumption with a given delay constraint. It is also worth noting that for each value ofw1, the policy learned from ECOO always has better performance in terms of both power consumption and buffering delay, which demonstrates the superiority of the other based strategy for continuous power control.

    (19)

    The range of JFI values is [1/N,1], where JFI=1 when all users have the exact same rate. Therefore, the closer JFI is 1, the better the fairness between users will be proved. As shown in Fig.11, the JFI of the proposed ECOO is about 0.9, which is much higher than that of the pre-optimization scheme. Moreover, with the given number of cellular users, this advantage will be any more evident as the number of users increases.

    Fig.11 Jain’s fairness index

    5 Conclusions

    A multi-user MEC system is designed, and the system sets conditions such as the random arrival of tasks and the change of wireless channels in each user with time. In order to minimize the power consumption and buffering delay, a decentralized dynamic task is designed based on the DRL uninstall algorithm, and the ECOO algorithm has been applied successfully to every mobile user offloading of autonomic learning strategy. The strategy can be observed from MEC local system according to the result of adaptive allocation of local computing tasks or missions offloading. Experimental simulations show that in a decentralized strategy, ECOO has better simulation results in terms of latency and energy consumption than traditional DQN and DDPG algorithms. By analyzing the power delay trade-off of each user, it was found that ECOO also had better results. However, the main consideration in this article is the slower channel changes caused by mobility, the system capacity computation does not take into account fast fading effects and thus will not reflect real capacity performance of the network. In the further research work, it is expected to base resource management on slow fading parameters and statistical information of the channel instead of instantaneous CSI to address the challenges caused by the inability to track fast changing wireless channels, and further optimize the ECOO algorithm.

    中亚洲国语对白在线视频| 久久精品国产综合久久久| 国产成人精品无人区| 两个人免费观看高清视频| 亚洲欧美日韩另类电影网站| 我的亚洲天堂| 国产男女内射视频| 国产97色在线日韩免费| 后天国语完整版免费观看| 天天添夜夜摸| 国产在线视频一区二区| 少妇被粗大的猛进出69影院| 精品人妻1区二区| 色婷婷久久久亚洲欧美| 色视频在线一区二区三区| 国产区一区二久久| 桃红色精品国产亚洲av| 国产有黄有色有爽视频| 人人妻人人爽人人添夜夜欢视频| 亚洲欧美精品综合一区二区三区| 91麻豆av在线| 操出白浆在线播放| a级毛片在线看网站| 女人久久www免费人成看片| videos熟女内射| 欧美精品亚洲一区二区| 老司机在亚洲福利影院| 亚洲成国产人片在线观看| 亚洲黑人精品在线| 亚洲七黄色美女视频| 国产三级黄色录像| 午夜福利视频在线观看免费| 国产精品久久久久久精品古装| a 毛片基地| 免费在线观看视频国产中文字幕亚洲 | 老司机影院毛片| 亚洲第一青青草原| 国产又爽黄色视频| 午夜影院在线不卡| 韩国精品一区二区三区| 亚洲三区欧美一区| 婷婷色av中文字幕| 欧美日韩黄片免| 岛国在线观看网站| 免费在线观看视频国产中文字幕亚洲 | 国产成人免费无遮挡视频| 成人18禁高潮啪啪吃奶动态图| 日本一区二区免费在线视频| 成年人黄色毛片网站| 国产精品久久久久久人妻精品电影 | 脱女人内裤的视频| 欧美另类一区| 最新的欧美精品一区二区| 日韩人妻精品一区2区三区| 亚洲九九香蕉| 精品国产一区二区三区久久久樱花| 日韩熟女老妇一区二区性免费视频| 国产欧美亚洲国产| 精品国产一区二区三区四区第35| 久久国产精品男人的天堂亚洲| 亚洲成人免费av在线播放| 在线观看www视频免费| 97精品久久久久久久久久精品| 欧美激情久久久久久爽电影 | 在线十欧美十亚洲十日本专区| 在线观看免费高清a一片| 两个人看的免费小视频| 国产亚洲午夜精品一区二区久久| 欧美成狂野欧美在线观看| 在线观看免费日韩欧美大片| 黄频高清免费视频| 最近中文字幕2019免费版| 80岁老熟妇乱子伦牲交| 国产亚洲精品第一综合不卡| 国产精品国产三级国产专区5o| 亚洲av成人一区二区三| 99久久国产精品久久久| 在线天堂中文资源库| 亚洲少妇的诱惑av| 一级毛片电影观看| 永久免费av网站大全| 亚洲男人天堂网一区| 丁香六月天网| 欧美激情久久久久久爽电影 | 日韩欧美国产一区二区入口| 在线 av 中文字幕| 热99国产精品久久久久久7| 王馨瑶露胸无遮挡在线观看| 男男h啪啪无遮挡| 色播在线永久视频| 中文字幕高清在线视频| 一级毛片女人18水好多| 男女高潮啪啪啪动态图| 一级片'在线观看视频| 国产亚洲欧美在线一区二区| 麻豆乱淫一区二区| 美女视频免费永久观看网站| 日韩 亚洲 欧美在线| av免费在线观看网站| 曰老女人黄片| 丝袜脚勾引网站| 欧美精品亚洲一区二区| 国产精品1区2区在线观看. | av在线播放精品| 亚洲色图综合在线观看| 中文字幕av电影在线播放| 亚洲中文日韩欧美视频| 精品国产乱子伦一区二区三区 | 亚洲五月色婷婷综合| 97精品久久久久久久久久精品| 黑人巨大精品欧美一区二区mp4| 久久国产精品影院| 亚洲精品成人av观看孕妇| 99热全是精品| 国产日韩欧美视频二区| 九色亚洲精品在线播放| 水蜜桃什么品种好| 黑人巨大精品欧美一区二区mp4| 12—13女人毛片做爰片一| 又黄又粗又硬又大视频| 丁香六月欧美| 高清在线国产一区| 国产精品av久久久久免费| 国产精品.久久久| 精品亚洲成a人片在线观看| 欧美精品高潮呻吟av久久| 日本撒尿小便嘘嘘汇集6| 18在线观看网站| 亚洲伊人久久精品综合| 午夜福利在线观看吧| 在线观看人妻少妇| 免费久久久久久久精品成人欧美视频| 欧美xxⅹ黑人| 男女免费视频国产| 黄色视频不卡| 欧美日韩亚洲高清精品| 国产成人精品久久二区二区免费| 亚洲欧洲精品一区二区精品久久久| 久久久精品94久久精品| 国产亚洲欧美精品永久| 亚洲伊人久久精品综合| 精品国产一区二区久久| 青草久久国产| 欧美变态另类bdsm刘玥| 青春草视频在线免费观看| 男人舔女人的私密视频| 国产高清视频在线播放一区 | 久久99热这里只频精品6学生| 中文字幕最新亚洲高清| 男女高潮啪啪啪动态图| 国产成人啪精品午夜网站| 国产成人啪精品午夜网站| 法律面前人人平等表现在哪些方面 | 99热全是精品| 久久久精品区二区三区| 亚洲三区欧美一区| 亚洲av片天天在线观看| 午夜91福利影院| 午夜久久久在线观看| 日本撒尿小便嘘嘘汇集6| 国产成人一区二区三区免费视频网站| 成人黄色视频免费在线看| 如日韩欧美国产精品一区二区三区| 久久人妻福利社区极品人妻图片| 久久av网站| 99re6热这里在线精品视频| 丁香六月天网| 12—13女人毛片做爰片一| 美女国产高潮福利片在线看| 国产精品1区2区在线观看. | 日韩中文字幕视频在线看片| 99热国产这里只有精品6| 欧美日韩成人在线一区二区| 成人黄色视频免费在线看| 国产欧美日韩综合在线一区二区| 亚洲欧美成人综合另类久久久| 青春草亚洲视频在线观看| 乱人伦中国视频| 精品少妇一区二区三区视频日本电影| 女人久久www免费人成看片| 免费观看a级毛片全部| 精品一区二区三卡| 亚洲av电影在线观看一区二区三区| av欧美777| 国产日韩欧美亚洲二区| 免费少妇av软件| 99精品欧美一区二区三区四区| 国产在线视频一区二区| 亚洲国产精品999| 黄片小视频在线播放| 亚洲av电影在线进入| 51午夜福利影视在线观看| 啪啪无遮挡十八禁网站| 国产熟女午夜一区二区三区| 精品免费久久久久久久清纯 | 欧美大码av| 伦理电影免费视频| 香蕉国产在线看| 嫁个100分男人电影在线观看| 视频在线观看一区二区三区| 国产黄色免费在线视频| 精品国产乱码久久久久久男人| 老熟妇乱子伦视频在线观看 | avwww免费| 秋霞在线观看毛片| 人妻久久中文字幕网| 国产日韩一区二区三区精品不卡| 精品高清国产在线一区| 免费久久久久久久精品成人欧美视频| 黑人猛操日本美女一级片| 男女高潮啪啪啪动态图| 老司机午夜十八禁免费视频| 大型av网站在线播放| 狠狠狠狠99中文字幕| 欧美日韩成人在线一区二区| 久9热在线精品视频| 国产一区二区在线观看av| 婷婷丁香在线五月| 性色av乱码一区二区三区2| 性色av乱码一区二区三区2| 亚洲精品自拍成人| 日韩电影二区| 男人爽女人下面视频在线观看| 欧美黄色淫秽网站| 老鸭窝网址在线观看| 51午夜福利影视在线观看| 天天添夜夜摸| 另类亚洲欧美激情| 美女福利国产在线| 欧美日韩福利视频一区二区| 久久国产精品男人的天堂亚洲| 99精品欧美一区二区三区四区| 99国产综合亚洲精品| 欧美亚洲 丝袜 人妻 在线| 午夜视频精品福利| 丰满少妇做爰视频| 亚洲三区欧美一区| 男人舔女人的私密视频| 少妇 在线观看| 嫩草影视91久久| 少妇精品久久久久久久| 乱人伦中国视频| 在线观看人妻少妇| 成年女人毛片免费观看观看9 | 91老司机精品| 亚洲精品av麻豆狂野| 9热在线视频观看99| 啦啦啦啦在线视频资源| 亚洲精品一卡2卡三卡4卡5卡 | av一本久久久久| 日韩大片免费观看网站| 国产精品 欧美亚洲| 亚洲avbb在线观看| 日本vs欧美在线观看视频| 亚洲精品国产精品久久久不卡| 亚洲欧美清纯卡通| 美女高潮喷水抽搐中文字幕| 国产免费现黄频在线看| 人人妻人人爽人人添夜夜欢视频| 国产高清视频在线播放一区 | 午夜影院在线不卡| 午夜福利,免费看| 亚洲精品美女久久av网站| 欧美97在线视频| 老司机靠b影院| 俄罗斯特黄特色一大片| 97精品久久久久久久久久精品| 精品久久久久久久毛片微露脸 | 久久中文字幕一级| 女人久久www免费人成看片| 每晚都被弄得嗷嗷叫到高潮| 午夜福利视频精品| 国产日韩一区二区三区精品不卡| 久久精品人人爽人人爽视色| 纵有疾风起免费观看全集完整版| 好男人电影高清在线观看| 国产精品熟女久久久久浪| 伊人亚洲综合成人网| 国产亚洲av高清不卡| 丰满迷人的少妇在线观看| 日本a在线网址| 国产深夜福利视频在线观看| 老熟女久久久| 自线自在国产av| 考比视频在线观看| 久久狼人影院| 日韩人妻精品一区2区三区| 国产精品一区二区免费欧美 | 亚洲精华国产精华精| 老司机影院毛片| 最新在线观看一区二区三区| a级毛片在线看网站| 亚洲综合色网址| 日韩三级视频一区二区三区| 国产片内射在线| 人人妻人人澡人人爽人人夜夜| 欧美人与性动交α欧美精品济南到| 精品人妻在线不人妻| 国产一区二区三区综合在线观看| 午夜福利视频精品| www日本在线高清视频| 啦啦啦中文免费视频观看日本| 黄色a级毛片大全视频| 最新在线观看一区二区三区| 久久影院123| 国产成人欧美| 精品人妻熟女毛片av久久网站| 国产精品久久久久久精品电影小说| 精品卡一卡二卡四卡免费| 久久热在线av| 亚洲情色 制服丝袜| 欧美一级毛片孕妇| 高清在线国产一区| 国产男人的电影天堂91| 精品少妇久久久久久888优播| 午夜91福利影院| 色老头精品视频在线观看| 久久综合国产亚洲精品| 亚洲五月婷婷丁香| 亚洲国产av影院在线观看| 国产一区二区三区综合在线观看| 12—13女人毛片做爰片一| 久久人人爽av亚洲精品天堂| 中文精品一卡2卡3卡4更新| 1024视频免费在线观看| 五月开心婷婷网| 青草久久国产| 老司机午夜福利在线观看视频 | 80岁老熟妇乱子伦牲交| 一区二区三区四区激情视频| 亚洲av男天堂| 国产老妇伦熟女老妇高清| 中文字幕人妻丝袜一区二区| 黄色片一级片一级黄色片| 国产在视频线精品| kizo精华| 亚洲免费av在线视频| 多毛熟女@视频| 日韩中文字幕欧美一区二区| 国产精品一区二区在线观看99| 老司机影院毛片| av片东京热男人的天堂| 丰满迷人的少妇在线观看| 美女中出高潮动态图| 麻豆国产av国片精品| 精品少妇久久久久久888优播| 99国产精品一区二区三区| 国产精品亚洲av一区麻豆| 九色亚洲精品在线播放| 日本五十路高清| 精品少妇黑人巨大在线播放| 天堂中文最新版在线下载| 国产人伦9x9x在线观看| 亚洲av成人不卡在线观看播放网 | 午夜福利视频精品| 欧美午夜高清在线| 美女主播在线视频| 女人被躁到高潮嗷嗷叫费观| 男女下面插进去视频免费观看| 午夜免费鲁丝| 亚洲国产看品久久| 国产在线免费精品| 女警被强在线播放| 99久久99久久久精品蜜桃| 亚洲,欧美精品.| 免费不卡黄色视频| 秋霞在线观看毛片| 美女中出高潮动态图| 深夜精品福利| a 毛片基地| 国产精品一区二区免费欧美 | 日韩欧美一区二区三区在线观看 | 人成视频在线观看免费观看| 黄色片一级片一级黄色片| av福利片在线| 免费一级毛片在线播放高清视频 | 亚洲精品国产精品久久久不卡| 最近最新中文字幕大全免费视频| 国产色视频综合| 热99国产精品久久久久久7| 免费一级毛片在线播放高清视频 | 女人高潮潮喷娇喘18禁视频| 日本a在线网址| 国产av国产精品国产| 狠狠精品人妻久久久久久综合| 国产精品久久久久久人妻精品电影 | 久久精品成人免费网站| 淫妇啪啪啪对白视频 | 国产精品久久久久久人妻精品电影 | 法律面前人人平等表现在哪些方面 | 国产有黄有色有爽视频| 黄片播放在线免费| 久久人人爽人人片av| 肉色欧美久久久久久久蜜桃| 伦理电影免费视频| 搡老熟女国产l中国老女人| 无遮挡黄片免费观看| 黄色片一级片一级黄色片| 成人国产av品久久久| 嫩草影视91久久| 欧美久久黑人一区二区| 女性被躁到高潮视频| 黄色视频在线播放观看不卡| 悠悠久久av| 国产欧美亚洲国产| 亚洲av欧美aⅴ国产| 中文欧美无线码| 亚洲国产精品一区三区| 亚洲成人免费电影在线观看| 精品人妻在线不人妻| 午夜福利在线观看吧| 国产日韩一区二区三区精品不卡| 亚洲va日本ⅴa欧美va伊人久久 | 精品一区在线观看国产| 伦理电影免费视频| 欧美激情久久久久久爽电影 | 9191精品国产免费久久| 亚洲精品国产区一区二| 男女之事视频高清在线观看| av天堂久久9| 欧美黑人精品巨大| 91麻豆av在线| 久久精品aⅴ一区二区三区四区| 欧美 亚洲 国产 日韩一| 新久久久久国产一级毛片| 操美女的视频在线观看| 丝袜喷水一区| 精品欧美一区二区三区在线| 欧美日韩亚洲高清精品| 国产成人精品在线电影| 国产av又大| 国产亚洲精品一区二区www | 午夜免费成人在线视频| 欧美变态另类bdsm刘玥| 这个男人来自地球电影免费观看| 在线观看舔阴道视频| 亚洲,欧美精品.| 亚洲成国产人片在线观看| 飞空精品影院首页| 成年人免费黄色播放视频| 老司机午夜福利在线观看视频 | 国产主播在线观看一区二区| 欧美少妇被猛烈插入视频| 两人在一起打扑克的视频| 欧美日韩精品网址| 欧美日韩亚洲综合一区二区三区_| 国产精品九九99| 人妻久久中文字幕网| 天堂俺去俺来也www色官网| 午夜福利免费观看在线| 精品一区二区三区四区五区乱码| 亚洲精品第二区| 男女下面插进去视频免费观看| 久久99热这里只频精品6学生| 久久久欧美国产精品| 岛国毛片在线播放| 欧美中文综合在线视频| 午夜免费观看性视频| 一区在线观看完整版| 两个人看的免费小视频| 午夜免费成人在线视频| 中文字幕人妻丝袜一区二区| 欧美人与性动交α欧美精品济南到| 一本综合久久免费| 最近中文字幕2019免费版| 欧美精品av麻豆av| 久久久精品免费免费高清| 国产黄频视频在线观看| 午夜日韩欧美国产| 免费在线观看完整版高清| 亚洲伊人久久精品综合| 免费在线观看视频国产中文字幕亚洲 | 天堂俺去俺来也www色官网| 成人免费观看视频高清| 操美女的视频在线观看| 欧美变态另类bdsm刘玥| 欧美另类一区| 国产精品.久久久| 夜夜夜夜夜久久久久| tube8黄色片| 成人影院久久| 一区福利在线观看| 国产99久久九九免费精品| 久久久久久免费高清国产稀缺| 99香蕉大伊视频| 黄色视频在线播放观看不卡| 99香蕉大伊视频| 国产野战对白在线观看| 亚洲一卡2卡3卡4卡5卡精品中文| 99热全是精品| 老司机影院毛片| 亚洲情色 制服丝袜| 动漫黄色视频在线观看| 婷婷丁香在线五月| 免费在线观看视频国产中文字幕亚洲 | 中文字幕色久视频| 亚洲专区中文字幕在线| 精品国产乱码久久久久久小说| 午夜福利视频在线观看免费| 一个人免费在线观看的高清视频 | 黑人操中国人逼视频| 蜜桃在线观看..| 精品久久久久久久毛片微露脸 | 亚洲综合色网址| 亚洲第一av免费看| 性少妇av在线| 亚洲伊人色综图| 99热国产这里只有精品6| 成年av动漫网址| 美女大奶头黄色视频| 久久精品国产综合久久久| 成人免费观看视频高清| 欧美精品av麻豆av| 国产91精品成人一区二区三区 | 久久女婷五月综合色啪小说| 免费观看a级毛片全部| 国产av国产精品国产| 亚洲第一av免费看| 狠狠狠狠99中文字幕| 国产成人一区二区三区免费视频网站| 一级黄色大片毛片| 色综合欧美亚洲国产小说| 一级毛片电影观看| 久久毛片免费看一区二区三区| 亚洲精品国产一区二区精华液| 秋霞在线观看毛片| 777久久人妻少妇嫩草av网站| 国产在线观看jvid| 男女之事视频高清在线观看| av福利片在线| 国产亚洲精品一区二区www | 国产高清国产精品国产三级| 777米奇影视久久| 国产精品免费视频内射| 亚洲精品自拍成人| 国产成人免费无遮挡视频| www.av在线官网国产| 亚洲 欧美一区二区三区| 亚洲熟女精品中文字幕| 午夜激情av网站| 天堂中文最新版在线下载| av网站在线播放免费| 1024香蕉在线观看| 美女高潮到喷水免费观看| 免费日韩欧美在线观看| 90打野战视频偷拍视频| 亚洲精品成人av观看孕妇| 操美女的视频在线观看| 午夜影院在线不卡| 国产成人欧美在线观看 | 午夜福利影视在线免费观看| 亚洲av美国av| 日本vs欧美在线观看视频| 99热国产这里只有精品6| 久久这里只有精品19| 老司机亚洲免费影院| 少妇被粗大的猛进出69影院| 两人在一起打扑克的视频| 国产无遮挡羞羞视频在线观看| 五月开心婷婷网| 亚洲免费av在线视频| 精品少妇久久久久久888优播| 美女视频免费永久观看网站| 亚洲av日韩在线播放| 黄片播放在线免费| 97精品久久久久久久久久精品| 黄色视频不卡| 久久99一区二区三区| 亚洲av成人一区二区三| 日韩欧美免费精品| 久久久久久亚洲精品国产蜜桃av| 黑人巨大精品欧美一区二区蜜桃| 国产伦人伦偷精品视频| 老司机深夜福利视频在线观看 | 80岁老熟妇乱子伦牲交| 国产极品粉嫩免费观看在线| 99久久99久久久精品蜜桃| 欧美黑人精品巨大| 国产成人啪精品午夜网站| 老熟妇乱子伦视频在线观看 | 视频区图区小说| 青春草亚洲视频在线观看| 男人添女人高潮全过程视频| 黄色片一级片一级黄色片| 亚洲精品自拍成人| 老司机深夜福利视频在线观看 | 99久久99久久久精品蜜桃| 国内毛片毛片毛片毛片毛片| 少妇的丰满在线观看| 久久久久久人人人人人| 岛国毛片在线播放| 一级毛片电影观看| kizo精华| 一本色道久久久久久精品综合| 亚洲欧美日韩另类电影网站| 国产伦人伦偷精品视频| 手机成人av网站| 欧美日韩视频精品一区| av在线app专区| 2018国产大陆天天弄谢| 精品一区二区三卡| 欧美中文综合在线视频| 午夜免费成人在线视频| 最新在线观看一区二区三区| 777久久人妻少妇嫩草av网站| 亚洲午夜精品一区,二区,三区| 80岁老熟妇乱子伦牲交| 精品一区二区三区四区五区乱码| 精品亚洲成a人片在线观看| 久久久国产成人免费| 无限看片的www在线观看| 成人免费观看视频高清| 精品少妇久久久久久888优播| 狂野欧美激情性xxxx| 久久久国产欧美日韩av|