• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Self-Attention Based Dynamic Resource Management for Satellite-Terrestrial Networks

    2024-04-28 11:59:28LinTianhaoLuoZhiyong
    China Communications 2024年4期

    Lin Tianhao ,Luo Zhiyong,2,*

    1 School of Electronics and Communication Engineering,Sun Yat-Sen University,Shenzhen 518107,China

    2 Peng Cheng Laboratory,Shenzhen 518107,China

    Abstract: The satellite-terrestrial networks possess the ability to transcend geographical constraints inherent in traditional communication networks,enabling global coverage and offering users ubiquitous computing power support,which is an important development direction of future communications.In this paper,we take into account a multi-scenario network model under the coverage of low earth orbit (LEO)satellite,which can provide computing resources to users in faraway areas to improve task processing effciiency.However,LEO satellites experience limitations in computing and communication resources and the channels are time-varying and complex,which makes the extraction of state information a daunting task.Therefore,we explore the dynamic resource management issue pertaining to joint computing,communication resource allocation and power control for multi-access edge computing(MEC).In order to tackle this formidable issue,we undertake the task of transforming the issue into a Markov decision process(MDP)problem and propose the self-attention based dynamic resource management(SABDRM)algorithm,which effectively extracts state information features to enhance the training process.Simulation results show that the proposed algorithm is capable of effectively reducing the long-term average delay and energy consumption of the tasks.

    Keywords: mobile edge computing;resource management;satellite-terrestrial networks;self-attention

    I.INTRODUCTION

    With the emergence of computation-intensive applications such as Internet of Remote Things (IoRT),military training and geological exploration,traditional terrestrial networks are increasingly unable to fulflil the requirements of ubiquitous transmission.This is primarily due to the inability to provide coverage in remote areas such as mountains,oceans,and deserts[1].Additionally,ground-based communication infrastructure is susceptible to factors such as typhoons,earthquakes,and conflicts.As a result,there is an increasing urgency for wireless networks that can provide long-term seamless coverage.Recently,signifciant developments in satellite-terrestrial networks have occurred,as seen by programs such as Starlink,which entail the deployment of numerous satellites to provide a wide range of services to commercial,civil,and military customers.Technology companies such as OneWeb[2],SpaceX[3],O3b[4],and TeleSat[5]have been actively launching satellites to offer communication services.Satellite communication offers several advantages over terrestrial networks,including more flexible network access services,extensive coverage,and enhanced capacity for backhaul [6].Notably,the utilization of low earth orbit(LEO)satellites equipped with edge servers enables them to serve as wireless edge access nodes within satellite-terrestrial networks,delivering both computing power and communication support to users.

    On another front,the extensive advancement of hardware has resulted in the signifciant development of time-sensitive applications,including real-time gaming,intelligent transportation,high-defniition streaming media,and voice recognition [7].This has created a demand for satellites to not only offer communication services with global coverage but also services with low delay and high quality[8,9].Considering the limited resources of satellites,managing resources for computing and communication effciiently and reasonably is necessary [10].To reduce delay,users frequently prefer to offload tasks to the edge servers on LEO satellites rather than cloud servers.Wanget al.[11]presented a double edge computation offloading method for satellite-terrestrial networks to optimize latency and energy consumption,where terrestrial base stations and LEO satellites serve as edge nodes.Fanget al.[12] proposed a scheme based on many-to-one matching theory and game-theoretic to address the joint communication and computing resource allocation problem in satellite networks.In[13],the Orbital Edge Computing(OEC)task allocation algorithm was studied to reduce the computational cost.Authors in[14]introduced an effective computation architecture and proposed a distributed algorithm that utilizes the alternating direction method of multipliers (ADMMs) to effectively reduce the overall energy consumption of ground users.In [15],the minimization problem of the weighted sum energy consumption for terrestrial-satellite networks was decomposed into two layered subproblems and an energy effciient algorithm was proposed.However,most of the existing works employ conventional optimization algorithms to solve joint optimization problems,which necessitate numerous iterative solutions or cannot be generalized to complex scenarios.Thus,real-time guarantees cannot be ensured due to the algorithm’s substantial computational workload.

    With this background,we consider a multi-scenario network model under the coverage of the LEO satellite with artifciial intelligence in this article,where users in different regions can offload tasks to the LEO satellite equipped with multi-access edge computing(MEC) server in order to minimize the energy consumption and delay of tasks.Although some deep reinforcement learning-based algorithms are proposed in [16-19],complex and time-varying satellite channel state information is still diffciult to analyze and extract features effciiently.To tackle this,we propose a self-attention based resource management algorithm,which can effectively enhance the feature extraction and learning process.To the best of our knowledge,this is the frist work that applies self-attention to the resource management problem in satellite-terrestrial networks.According to[20],the self-attention mechanism can capture the long-range dependencies and semantic relationships among the input vectors,and has been widely used in natural language processing and object detection.Moreover,it is effective in analyzing the state characteristics and requires signifciantly less time to train.By incorporating self-attention into the policy network and the Q network of the deep reinforcement learning framework,we can achieve superior performance in terms of delay and energy consumption.In general,we transform the original issue into a Markov decision process (MDP) problem and utilize a self-attention based resource management algorithm to jointly optimize the computing and communication resource allocation as well as power control.The key features of this paper are as follows:

    ? We propose a self-attention based resource management algorithm that jointly optimizes computing,communication resource allocation,and power control for satellite-terrestrial networks.This algorithm addresses the long-term delay and energy consumption issue caused by the uneven distribution of computing resources on the satellite to different users.

    ? Different from other algorithms based on deep reinforcement learning,the proposed self-attention based dynamic resource management (SABDRM) algorithm incorporates a self-attention mechanism into reinforcement learning.This mechanism is capable of effciiently extracting state features to expedite training and effectively address the MDP problem in high-dimensional continuous action and state spaces.

    ? According to the simulation experiments,SABDRM outperforms other baseline algorithms in terms of delay and energy consumption performance.

    The remainder of this article is organized as follows.Section II introduces the system model and problem formulation.In Section III the SABDRM algorithm is proposed.Section IV discusses the simulation results.Finally,Section V concludes the article.

    II.SYSTEM MODEL AND PROBLEM FORMULATION

    The system model is represented in Figure 1 as a LEO satellite with a MEC server andNmobile users.We consider that users can offload tasks to the satellite in scenarios such as forests,deserts,cities,etc.There are two offloading modes for the users,namely local offloading and offloading to the LEO satellite.Based on the high-performance DRL-based mobile edge offloading scheme proposed in our previous work [21],we focus on resource management for each LEO satellite in this paper.Therefore,we only consider mobile users who offload their tasks to the satellite,presented byN={1,2,···,N}.Moreover,we useT={1,2,···,T}to denote all time slots,where each time slotthas durationτ[22].At each time slot,useri(i ∈N)generates a taskIi(t)={Li(t),Xi(t),Yi(t)},whereLi(t)denotes the computation amount of task,i.e.,the necessary CPU frequency required to fniish the task,andXi(t)represents the data size of the task[23].We denote the priority of the task asYi(t) andYi(t)∈Yis specifeid withY=[1,2,···,PN].The smaller the number,the higher the priority.The main notations used in this article are listed in Table 1.

    Table 1.List of notations.

    Figure 1.System model of multi-scenario communication under MEC-enabled satellite-terrestrial networks.

    2.1 Communication Model

    We assume that users adopt frequency division multiple access(FDMA)for their task offloading so that all users share the spectrum resources with a bandwidth ofB.According to the Shannon formula,the uplink rate from mobile userito the LEO satellite in time slottis:

    whereβi(t)denotes the proportion of bandwidth allocated by LEO satellite to useriin time slott,piindicates the transmission power of useri,N0represents the Gaussian noise power spectral density.Considering shadowed-Rician fading as well as large-scale fading [24],the channel gain between the LEO satellite and useriis represented asgi.Specifcially,we havegi=G|li(t)|-ε,whereli(t) is the distance between LEO satellite and useri,Grepresents the antenna gain of useri,andεdenotes the path loss exponent[25].

    Since the data size processed after the task is relatively much smaller than the size of input computation data,we do not consider the delay and energy consumption in downloading [26].Typical service types include statistics,machine learning,and data mining.Also,letcdenote the speed of light in vacuum,then the transmission delay of the task offloading to the MEC server is

    The energy consumption associated with each user upload must not surpass the maximum energy capacity of its respective battery,so it should meet

    2.2 Computation Model

    Due to the users with different priorities accessing the LEO satellite and limited caching and computing resources on the LEO satellite,corresponding queuing delays may occur.Therefore,it is crucial to consider both the processing delay and the queuing delay of the task when a user offloads a task to the MEC server.For the edge computation,the computation delay consists of the processing delay and queueing delay.According to the above-mentioned,we defnie the CPU frequency of the LEO satellite asFand denote the computation delay of taskIion the LEO satellite as

    whereκi(t) represents the proportion of computational resources allocated to taskIiby LEO satellite in time slott.When calculating the queuing delay,we introduce knowledge about queuing theory.Queuing theory has been extensively utilized in the feilds of communication and computing[27].There are several task queuesM(t)=[M1(t),M2(t),...,MN(t)]in the MEC server on the LEO satellite to store the arriving tasks,whereMi(t) denotes the computation tasks generated by useriexisting in the queue at time slott.It is assumed that the arrival rate of tasks of any priority at any time slot to the queues follows a Poisson distribution with parameterλi(t)[28],and the processing time of the MEC server follows an exponential distribution with parameterμ(t).Taking into account the accumulation of tasks in each queue caused by unfniished tasks at each time slott,it is necessary to consider the situation after the priority tasks arrive.To proceed,we assume that each non-preemptive f-i nite queue in the LEO satellite is an M/M/1 queue and tasks of the same priority are processed obeying the frist-come-frist-served principle.Furthermore,the MEC server has suffciient storage space to ensure that tasks will not be discarded by the queues.Based on the proposed model,there existPNpriorities,which will influence the total arrival rate as follows

    Then the average queuing delay for taskIicorresponding to priorityYi(t)is

    The remaining parameters are as follows

    2.3 Problem Formulation

    Based on the previous study,it can be clearly known that the overall delay for all computation tasks includes the delay of transmitting computing tasks to the LEO satellite and the computation delay of tasks on the LEO satellite.We choose the average energy consumption and the average delay of a task as the performance metrics in the dynamic resource management[29].The average cost function for all computation tasks in time slottcould be calculated according to(11),whereσ ∈R+is the cost coeffciient.

    In edge computing systems,users have high requirements for both delay and energy consumption performance.Time-sensitive tasks with high priority demand lower delay,while users with small battery capacity require higher energy consumption performance on the users’ side.However,in practical scenarios,when the task with a large data size needs offloading to reduce the computation delay,the energy consumption as well as the transmission delay increase.To some extent,this means that reducing the computation delay of task results in increased transmission delay and energy consumption.Therefore,the cost function is formulated as (11) which trades off different performance measures by using the cost coeffciient.It can be found that the average cost can be derived as a function of transmission delay and computation delay.The cost to energy consumption and transmission delay ratio rises as the cost coeffciient increases,which has a greater impact on the system’s average cost.The optimization objective is to balance various performance measures based on users’requirements.

    In order to solve the resource management problem of satellite-terrestrial networks with time-varying wireless channel states and randomly arriving tasks with different priorities,based on the aforementioned system model,we establish a joint computing,communication resource allocation and power control problem to optimize the total average cost for users.Specifcially,different tasks are processed in parallel within each time slot,but the delay and energy consumption between various tasks affect each other.Therefore,we propose the optimization objective:

    The objective function (12) is to minimize the average cost in resource management over a period ofT.The frist constraint (13) limits the transmission power of useriduring time slott.Moreover,pminrepresents the minimum allowable transmission power,whilepmaxdenotes the maximum allowable transmission power.The second constraint(14) indicates that the energy consumption associated with each user upload can not surpass the maximum energy capacity of its respective battery.The third constraint (15) indicates the total amount of spectrum resources allotted to each user cannot exceed the total available bandwidth.The last constraint (16) states that the amount of computing resources allotted to different users by the LEO satellite cannot surpass the maximum amount of computing resources that the server itself can provide.The optimization problem can be formulated as a typical dynamic programming problem.In order to address this issue,we suggest making it a MDP framework,considering a fniite time horizonT.The solution method will be elaborated in the following subsection.

    III.PROPOSED SELF-ATTENTION BASED SCHEME

    In this section,we convert the problem (12) into a MDP problem and devise a self-attention based algorithm to solve it[16].

    3.1 MDP-Based Dynamic Resource Management Problem

    Suppose a MDP problem is a 4-element tupleM={S,A,P,R},whereSrepresents the state space,Adenotes the action space,P:S×A×S→[0,1]expresses the transition probability among all states,andR:S×A→R stands for the reward function.In every time slot,the agent will make a corresponding actiona(t)∈Abased on the current environments(t)∈S.The environment will then be updated tos(t+1)∈S,and the agent will receive a rewardr(t)from it for multiple cycles.Herein,we construct the MDP problem model in detail based on the above description.

    (1) State Space: The agent receives the state informations(t)from the satellite communication environment at each time slott.

    whereL(t)represents the amount of computation for all tasks,X(t)denotes the amount of data for all tasks,Y(t)stands for the priority of all tasks.E(t)andg(t)express the remaining battery energy information and wireless channel gain information of all users in time slott,respectively.

    (2)Action Space: The agent makes the corresponding actiona(t)based on the currently observed information about the environments(t)every time slott.

    wherep(t) expresses the users’ transmission power.β(t) andκ(t) indicate the proportion of bandwidth and computational resources allocated by the LEO satellite to all users at the current time slot,respectively.Specifcially,constraints (13),(14),and (16)specify the actions ina(t).

    (3) State Transition Probability: The probability distribution ofs(t+1) givens(t) and the selecteda(t) are indicated by the state transition probabilityP(s(t+1)|s(t),a(t)).The probability distribution is only determined by the environment,and the agent does not have any prior knowledge of it[30].

    (4)Reward Function: We defnier(t)as the reward function for performing actiona(t)in states(t).Generally,the MDP problem entails maximizing the cumulative reward function.However,our goal is to minimize the average cost from the optimization objective function(12)in this chapter.Therefore,based on the problem(12),we choose the opposite of cost as the reward function.Additionally,we defnie different reward functions depending on the constraints.At the time slott,the corresponding reward function is

    When the restriction (13)-(16) are satisfeid,the reward function is the average cost under the time slott.If it is not satisfeid,we set the reward function to an extremely small value[16].Since the test periodTis generally large,it is guaranteed that-C(t)?-TC(t)holds,making the model setup more scientifci.After converting the problem into a MDP problem,we will raise the self-attention based Soft Actor-Critic (SAC)framework.

    3.2 Self-Attention Based SAC Framework

    In the context of real-time resource management problems in large-scale user scenarios,the relationship between the inputs and the outputs of the environment state could be highly complicated,and extracting multidimensional features can be challenging.Traditional optimization methods often require a signifciant number of iterations,leading to the diffciulty of solving the real-time problem.In this paper,we propose a modifciation of the SAC based on self-attention as the underlying framework[31-33].In our module,there are three different types of neural networks: Policy network,Q network and Target network,the latter two of which are twins.As shown in Figure 2,they all use self-attention networks as nonlinear function approximators.Policy network takes the environment states(t) as its input and outputs actionsa(t) to the environment and Q networks.Actionsa(t) are the input to the Q network,which produces the soft Bell-Man residualto Policy network.Target network is used to slowly track the actual soft Q-function to improve stability.

    Figure 2.Self-attention based SAC framework.

    3.2.1 SAC Framework

    Reinforcement learning is usually used to solve tasks where the action and environment interact frequently,and SAC is a remarkable one to address both discrete and continuous control problems.In contrast to the conventional Proximal Policy Optimization(PPO)[34],SAC operates in an off-policy manner.This implies that it remains effective when faced with signifciant policy disparities,and offers enhanced sampling effciiency,allowing for the reuse of data generated from previous policies.Compared to Deep Deterministic Policy Gradient (DDPG) [35],SAC exhibits less sensitivity to hyper-parameters,rendering it more amenable to adjustments.Stochastic policy in SAC is accomplished through the utilization of maximum entropy which ensures that no valuable action or trajectory goes unnoticed and safeguards comprehensiveness in decision-making.SAC has a strong anti-interference ability and stable performance,using maximum entropy objective to develop a learning policy for increasingly diffciult work:

    whereαis a hyper-parameter representing the temperature coeffciient,which is often utilized to fnie-tune the entropy value.In general,we seek a learnable policy moduleπ:S×A →R,which is able to direct the agent in selecting appropriate actions.And we employρπto represent the state-action marginals of the trajectory distribution brought on by a policyπ.Given a policyπand a states(t),we get the actionvalue function,also known as Q-functionQ(st,at)=to capture the anticipated cumulative reward when actiona(t)is chosen in states(t)according to the policyπ,whereγ ∈[0,1] is the discount factor.

    In practice,action space typically contains vast continuous domains,necessitating the development of a realistic approximation to soft policy iteration.To that end,the algorithm constructs function approximation,which can be modeled as neural networksπ?(at|st)andQθ(st,at) to describe policyπand action-state value functionQ(st,at).The actor-network and the critic-network are updated by evenly sampling the mini-batch from the replay buffer during each time step.In order to improve the stability and robustness of learning,SAC also utilizes a target network forQ(st,at) and policyπto perform soft updates based on the learned network.In the policy improvement step,the information projection described by the Kullback-Leibler divergence is favorable.To put it another way,we update the policy for each state during the policy improvement step according to

    where?trepresents an input noise vector,which is sampled from a predetermined distribution,such as a spherical Gaussian.The loss function for the actornetworks can be formulated as:

    The gradient is as follows based on the mini-batch sample data:

    Besides the actor-networks,the loss function for the critic-networks that can be trained to minimize the soft Bellman residual is:

    The update incorporates a target soft Q-function,where the soft Q-function weights are averaged exponentially to determine the parameterswhich helps improve training stability.Furthermore,we parameterize two soft Q-functions with parametersθiand train them independently to optimizeJQ(θi) in order to eliminate positive bias during the policy improvement step.So we can approximate the gradient of(25)with

    Along with the soft Q-function and the policy,we use the following formula to calculate gradients forα:

    3.2.2 Self-Attention Mechanism

    The preceding discussion focuses on the SAC framework.To provide a clearer explanation of our proposed algorithm,we will now introduce the multi-head self-attention mechanism [20].In the feild of machine learning,the attention mechanism is popularly used.In comparison to traditional DNN and CNN models,the attention mechanism not only reduces the number of model parameters but also enhances feature extraction capabilities.Furthermore,the attention mechanism enables the handling of variable-length sequences and multimodal information,enhancing the overall model’s flexibility and versatility.Specifcially,the attention mechanism model initially transforms the input into matricesQ,K,andVrepresenting a set of queries,keys and values,respectively.The output of the scaled dot-product attention can be formulated as:

    wheredkis the dimension of queries and keys.Another benefciial way called multi-head attention is to linearly project the queries,keys and valueshtimes with various,learned linear projections todk,dkanddvdimensions,respectively.We simultaneously execute the attention function on each of the projected versions of the queries,keys,and values.Laterdvdimensional output values can be produced.The fnial values are created by concatenating the output values and then projecting them once again.The multi-head attention mechanism enables the model to simultaneously focus on various subspaces of information at different positions.The output of the multi-head selfattention is:

    3.3 Practical Algorithm

    Based on the self-attention mechanism and SAC framework,we propose the SABDRM algorithm.SABDRM algorithm is an offline algorithm and utilizes the stochastic policy of maximum entropy to explore more possibilities.As an offline algorithm,SABDRM requires multiple episodes of iterations,each of which consists of an environmental step and a gradient step.At the end of the last episode,the agent will make a real decision.

    During each episode,the agent takes actiona(t)based on the current policyπ?(at|st) and the collected states(t) for each environment step.After getting an actiona(t),the system environment provides the agent with the next states(t+1) and rewardr(t),according to the state transition probabilityP(s(t+1)|s(t),a(t)).The agent then stores the transition in the replay buffer for further learning and decision-making.

    For each gradient step,two Q networks and the policy network update their model parametersθiand?by sampling mini-batches of dataN*(s(t),a(t),r(t),s(t+1)) from the replay buffer,respectively.After that the system updates the temperature coeffciientαand target network weightsBy iterative sampling and updating the model parameters with the replay buffer,the policy network and the Q networks gradually improve their performance and learn to make better decisions in the given environment.

    In contrast to conventional deep reinforcement learning,the policy network and Q networks in SABDRM employ a network with the multi-head selfattention mechanism rather than DNN.This facilitates enhanced extraction of features from the input vectors,as demonstrated in Section IV through experimental simulations.

    The fnial algorithm is presented in Algorithm 1.In practice,we take one gradient step after a single environment step.The approach involves an iterative process of gathering experience from the environment and updating the function approximators.Additionally,the former uses the current policy and the stochastic gradients from batches sampled from a replay buffer are utilized by the latter.

    IV.SIMULATION RESULTS

    In this section,we conduct a series of simulation experiments to evaluate the performance of the SABDRM algorithm in satellite-terrestrial networks,which is shown in Figure 3 to Figure 9.To facilitate the comparison,the average cost is applied in the baselines including the DDPG scheme,random resource management (RRM) scheme and uniform resource management(URM)scheme.

    Figure 3.The convergence curve of different schemes.

    ? DDPG scheme: The deep reinforcement learning algorithm DDPG is an actor-critic,modelfree algorithm that operates over continuous action spaces and is based on the deterministic policy gradient.

    ? RRM scheme: Its basic thesis is that the MEC server randomly assigns the spectrum resource and the computing resource to each task.Moreover,each user’s transmission power is random.

    ? URM scheme: For URM,the MEC server evenly allocates computing resources and spectrum resources to each user.The transmission power of each user is the average of the upper and lower power limits.

    We compare the average cost of the proposed SABDRM algorithm with the three baselines in terms of the computation amount of task,the data size of task,the number of priority levels of task,the total computing capacity of the LEO satellite and the number of users.

    4.1 Experiment Setup

    Simulation is performed on Python 3.9.We set up the LEO satellite with an altitude of 784 km flying across a square area of 1200 m × 1200 m [36],and ground users with different priorities randomly distributed.In addition,the channel information can be obtained in advance by the sensing technique.

    In the simulation,we set the channel bandwidth to 20 MHz with the Gaussian noise power spectrum densityN0=-174 dBm/Hz,and the channel model is chosen as shadowed-Rician fading as well as largescale fading channel model mentioned above.We mainly consider the computation amount,data size,priority,and possible energy consumption for the task parameters.The minimum limit of transmission powerpminis 23 dBm and the maximum limitpmaxis 38 dBm,according to[37].Besides,we assume a user’s maximum battery capacity is 1000 J[16]and the totalcomputing capacity of the LEO satellite is taken fromU[15,35]GHz.

    In the SABDRM algorithm,we use the multi-head self-attention network instead of DNN as the feature extraction layer of deep reinforcement learning followed by a flatten layer.What remains constant is that both the input and output layers are DNNs.After careful adjustment,we employh=8 heads.The characteristic dimension of the output is 256,and the dropout rate is 0.The training parameters of the model have a discount of 0.99.In addition,the main parameter settings are detailed in Table 2.

    Table 2.Experiments parameters.

    4.2 Performance and Analysis

    Figure 3 shows the convergence curves of different reinforcement learning algorithms.The learning rate is set as 10-3.It can be seen in Figure 3 that the proposed SABDRM has better performance than DDPG.Specifcially,the average cost of SABDRM can obtain 3 after 50 episodes.In contrast,the average cost of DDPG fluctuates around 5 after 175 episodes.This comparison proves that the benefti of the selfattention mechanism can better extract state features to speed up training and solve the MDP problem in high-dimensional continuous action and state spaces effectively.

    Figure 4 illustrates the average cost of users for the four algorithms with varying computation amounts of task.The number of priority levels of task and the number of users are set to 3 and 10,respectively.The data size of task and the total computing capacity in the LEO satellite follow Gaussian distributions with mean values of 3.5 MB and 25 GHz,respectively.Figure 4 denotes that the average task cost increases as the computation amount increases.The underlying explanation lies in that,with a fxied total computation capacity on the satellite,an increased computational load per offloaded task results in elevated delay and energy consumption.Consequently,this leads to a higher average cost across the entire system.Moreover,the average cost of the SABDRM is lower than the counterparts of the other three baselines.In particular,SABDRM effectively mitigates cost by 53.7%,48.6%,and 29.6% when compared to RRM,URM,and DDPG,respectively.

    Figure 4.Average cost of ground users versus different computation amount of task.

    Figure 5 depicts the average cost of task for URM,RRM,DDPG and SABDRM versus different data sizes.Here the number of priority levels of task and the number of users are set to 3 and 10,respectively.The computation amount of task and the total computing capacity in the LEO satellite follow Gaussian distributions with mean values of 3.5 Gcycles and 25 GHz,respectively.It can be observed from Figure 5 that the average cost of task increases with the data size.According to equation(2)-(4),the maximum battery capacity of the users limits the transmission energy consumption.Consequently,the average delay increases along with the data size of task,resulting in a rise in average cost.Furthermore,the proposed SABDRM performs better than RRM,URM and DDPG,which can effectively lower the average cost of task by 54.3%,46.4%,and 26.6%,respectively.

    Figure 5.Average cost of ground users versus different data size of task.

    Figure 6 shows the average cost of task the four algorithms with different numbers of priority levels of task.The number of users in this fgiure is set to 10.The computation amount of task,the data size of task and the total computing capacity in the LEO satellite follow Normal distributions with mean values of 3.5 Gcycles,3.5 MB and 25 GHz,respectively.The larger number of priority levels of task leads to higher queuing delay,which can increase the average cost of task.In addition,on average,the proposed SABDRM algorithm can reduce the average cost of task by 52.5%,50.1%and 29.7%compared with RRM,URM,and DDPG,respectively.The decisions created by RRM exhibit substantial uncertainties and are likely not optimal because of the randomness of RRM.Consequently,the trajectory of RRM tends to fluctuate.When the number of priority levels of task is large,URM will bring greater cost.The reason is that the impact of task priorities is ignored,which results in serious queuing delay.

    Figure 6.Average cost of ground users versus different number of priority levels of task.

    In Figure 7,we frist conduct a comparative analysis of the cost performance achieved by different resource management algorithms under varying total computing capacities of the LEO satellite.Each point in the fgiure represents the average performance across 1,000 independent wireless channel realizations [17].Here the number of priority levels of task and the number of users are set to 3 and 10,respectively.The data size and computation amount of task follow Gaussian distributions with mean values of 3.5 MB and 3.5 Gcycles,respectively.We see that the proposed SABDRM algorithm performs noticeably better than the other three baselines.In Figure 8,we further compare the cost performance of SABDRM and DDPG algorithms based on deep reinforcement learning.For better exposition,we plot both the median and the confdience intervals of average cost over 1,000 independent channel realizations.Specifcially,the central mark,highlighted in red,signifeis the median,and the bottom and top edges of the box represent the 25th and 75th percentiles,respectively.We fnid that the less the total computing capacity of the LEO satellite,the longer the average cost of the task.The increased total computing capacity signifeis that,on average,the LEO satellite possesses the ability to allocate a greater amount of computing power to each user,thus diminishing the average cost of task.Moreover,it can be observed that the median of the SABDRM algorithm is always lower than the counterpart of the DDPG algorithm and the upper and lower quartiles of SABDRM are spaced less than the DDPG,which means the deviation of the proposed SABDRM is smaller.As shown in Figure 7 and Figure 8,the SABDRM can reduce 32.1%of the average cost compared to the DDPG scheme.In contrast to the RRM and URM schemes,the SABDRM can reduce 54.8% and 48.6% of the average cost,respectively.

    Figure 7.Average cost of ground users versus different total computing capacity of the LEO satellite.

    Figure 8.Boxplot of the average cost of ground users for DDPG and SABDRM algorithms under different total computing capacity of the LEO satellite.

    Figure 9 fnially contrasts the average cost of users for the four algorithms versus various numbers of users.Here the number of priority levels of task is set to 3.The computation amount of task,the data size of task and the total computing capacity in the LEO satellite follow Normal distributions with mean values of 3.5 Gcycles,3.5 MB and 25 GHz,respectively.It can be observed that the average cost of task increases with the number of users.This phenomenon occurs due to the diminishing allocation of spectrum resources and computing power resources to each user as the number of users increases,consequently resulting in heightened delay and energy consumption.As shown in Figure 9,the SABDRM can reduce the average cost of task by 30.9%,17.4%and 10.1%compared with RRM,URM,and DDPG,respectively.

    Figure 9.Average cost of ground users versus different number of users.

    V.CONCLUSION

    In this paper,we have investigated the optimization problems of joint computing,communication resource allocation and power control between satelliteterrestrial networks.To address the dynamic and complex nature of satellite channels,the problem was converted to a MDP problem.Due to the large dimension of channel state in satellite-terrestrial networks and the challenge of feature extraction,we proposed the SABDRM algorithm based on the self-attention mechanism,which optimized the energy consumption and long-term average delay under the condition of limited computing power resources of LEO satellite and battery capacity of users.Extensive experimental results presented that,when compared to other baseline algorithms,our proposed algorithm showed superior performance in terms of reducing the average cost of users.Our work can offer valuable insights for future research on computing resource allocation for satellite-terrestrial networks and deep reinforcement learning algorithms.In the future,we will further explore collaborative resource management among satellites,effciient communication methods and mobility management in satellite-terrestrial networks,involving information theory and online optimization techniques.

    ACKNOWLEDGEMENT

    This work was supported by the National Key Research and Development Plan(No.2022YFB2902701)and the key Natural Science Foundation of Shenzhen(No.JCYJ20220818102209020).

    欧美激情 高清一区二区三区| 只有这里有精品99| 亚洲国产精品国产精品| 高清在线视频一区二区三区| 精品亚洲成a人片在线观看| 丁香六月天网| 日韩 欧美 亚洲 中文字幕| 飞空精品影院首页| 国产一级毛片在线| 十八禁高潮呻吟视频| 免费少妇av软件| 亚洲精品久久成人aⅴ小说| a级片在线免费高清观看视频| 亚洲视频免费观看视频| 成年人免费黄色播放视频| 大香蕉久久网| 亚洲精品国产av蜜桃| 操美女的视频在线观看| 免费高清在线观看视频在线观看| 久久99一区二区三区| 91精品三级在线观看| 久久久精品免费免费高清| 国产精品麻豆人妻色哟哟久久| 久久国产精品男人的天堂亚洲| 精品亚洲乱码少妇综合久久| 国产精品蜜桃在线观看| 久久精品久久久久久久性| 婷婷色av中文字幕| 18禁观看日本| 伊人久久国产一区二区| 精品人妻在线不人妻| 欧美人与性动交α欧美精品济南到| 十八禁高潮呻吟视频| 欧美成人精品欧美一级黄| 欧美老熟妇乱子伦牲交| 亚洲熟女精品中文字幕| 国产av国产精品国产| 另类精品久久| 99re6热这里在线精品视频| 成人手机av| 中文字幕最新亚洲高清| 啦啦啦啦在线视频资源| 欧美人与善性xxx| 亚洲美女黄色视频免费看| 性少妇av在线| 老鸭窝网址在线观看| 国产精品国产三级专区第一集| √禁漫天堂资源中文www| 美女国产高潮福利片在线看| 男人添女人高潮全过程视频| 交换朋友夫妻互换小说| 亚洲国产中文字幕在线视频| 人人妻人人添人人爽欧美一区卜| 极品人妻少妇av视频| 亚洲伊人久久精品综合| 别揉我奶头~嗯~啊~动态视频 | 狂野欧美激情性bbbbbb| 欧美日韩亚洲高清精品| 在线观看人妻少妇| 国产又色又爽无遮挡免| 久久国产精品大桥未久av| 精品国产露脸久久av麻豆| 人体艺术视频欧美日本| 欧美在线黄色| 久久国产精品男人的天堂亚洲| 十八禁高潮呻吟视频| 亚洲欧美成人精品一区二区| 高清视频免费观看一区二区| 亚洲国产中文字幕在线视频| 日韩制服骚丝袜av| 最新在线观看一区二区三区 | 麻豆av在线久日| 免费高清在线观看视频在线观看| 精品国产乱码久久久久久小说| √禁漫天堂资源中文www| 无限看片的www在线观看| 老司机靠b影院| 赤兔流量卡办理| 亚洲人成网站在线观看播放| 亚洲成人av在线免费| 亚洲精品久久久久久婷婷小说| 亚洲美女视频黄频| 国产一区二区在线观看av| 日韩制服骚丝袜av| 少妇人妻久久综合中文| 亚洲精品乱久久久久久| av天堂久久9| 99久久综合免费| 国产精品久久久久久精品电影小说| 午夜福利乱码中文字幕| 日本av免费视频播放| 天美传媒精品一区二区| 晚上一个人看的免费电影| 国产一区二区激情短视频 | 日韩大码丰满熟妇| 丁香六月欧美| 国产精品一二三区在线看| 亚洲图色成人| 久久久久久人人人人人| www.自偷自拍.com| 亚洲精品一二三| 最近中文字幕2019免费版| 国产黄色免费在线视频| 亚洲美女视频黄频| 亚洲av福利一区| 精品第一国产精品| 一区二区三区乱码不卡18| 婷婷成人精品国产| 欧美97在线视频| 十八禁网站网址无遮挡| 91精品三级在线观看| 中文字幕另类日韩欧美亚洲嫩草| 男的添女的下面高潮视频| 亚洲精品aⅴ在线观看| 七月丁香在线播放| 国产黄色视频一区二区在线观看| 亚洲国产看品久久| 黄片无遮挡物在线观看| 美女福利国产在线| 欧美97在线视频| 婷婷色麻豆天堂久久| 亚洲国产精品一区三区| 亚洲专区中文字幕在线 | 亚洲四区av| 亚洲精品久久成人aⅴ小说| 亚洲人成电影观看| 午夜福利视频精品| 美女午夜性视频免费| 天天躁夜夜躁狠狠久久av| 亚洲av电影在线观看一区二区三区| 国产精品久久久久久精品古装| 国产一区二区激情短视频 | 亚洲欧洲国产日韩| 色婷婷久久久亚洲欧美| 亚洲,欧美精品.| 婷婷色综合大香蕉| 在线 av 中文字幕| 日本黄色日本黄色录像| 日本vs欧美在线观看视频| 热99国产精品久久久久久7| 午夜免费男女啪啪视频观看| 日韩制服骚丝袜av| 国产精品免费大片| 亚洲av综合色区一区| 90打野战视频偷拍视频| 777米奇影视久久| 亚洲精品第二区| www.av在线官网国产| 亚洲免费av在线视频| 国产精品久久久久久精品电影小说| 黄片小视频在线播放| 亚洲一卡2卡3卡4卡5卡精品中文| 男人爽女人下面视频在线观看| 久久天堂一区二区三区四区| 色婷婷av一区二区三区视频| 下体分泌物呈黄色| 黄色视频不卡| 日韩av免费高清视频| 久久久精品区二区三区| 亚洲人成77777在线视频| 色综合欧美亚洲国产小说| 亚洲一区中文字幕在线| 精品少妇黑人巨大在线播放| 一本—道久久a久久精品蜜桃钙片| 欧美日韩精品网址| 色94色欧美一区二区| 桃花免费在线播放| 丰满饥渴人妻一区二区三| 成人亚洲精品一区在线观看| 中文字幕最新亚洲高清| 日韩制服丝袜自拍偷拍| 这个男人来自地球电影免费观看 | 在线天堂最新版资源| √禁漫天堂资源中文www| 亚洲国产欧美日韩在线播放| 中文天堂在线官网| 中文字幕av电影在线播放| 欧美精品高潮呻吟av久久| 18在线观看网站| 国产免费视频播放在线视频| 久久久精品免费免费高清| 少妇被粗大猛烈的视频| 精品少妇一区二区三区视频日本电影 | 巨乳人妻的诱惑在线观看| 亚洲人成电影观看| 亚洲自偷自拍图片 自拍| 亚洲第一青青草原| 在线观看三级黄色| 老熟女久久久| 日韩一本色道免费dvd| 国产av国产精品国产| 欧美在线黄色| 一区在线观看完整版| 99久国产av精品国产电影| 亚洲精品一二三| 99久久精品国产亚洲精品| 欧美精品亚洲一区二区| 19禁男女啪啪无遮挡网站| 色精品久久人妻99蜜桃| www.精华液| 女人被躁到高潮嗷嗷叫费观| 亚洲精品自拍成人| 精品久久久精品久久久| 欧美日本中文国产一区发布| 免费av中文字幕在线| 嫩草影视91久久| 日韩av不卡免费在线播放| 日日爽夜夜爽网站| 一区二区日韩欧美中文字幕| 国产精品亚洲av一区麻豆 | 亚洲伊人久久精品综合| 少妇 在线观看| 成人亚洲精品一区在线观看| 我的亚洲天堂| 国产精品麻豆人妻色哟哟久久| 欧美精品一区二区免费开放| 亚洲国产精品成人久久小说| 色网站视频免费| av一本久久久久| 高清在线视频一区二区三区| 哪个播放器可以免费观看大片| 欧美 亚洲 国产 日韩一| 久久精品国产亚洲av涩爱| av又黄又爽大尺度在线免费看| 蜜桃国产av成人99| 麻豆av在线久日| netflix在线观看网站| 亚洲精品一二三| 日韩大片免费观看网站| 18禁动态无遮挡网站| 永久免费av网站大全| 中文字幕av电影在线播放| 精品人妻在线不人妻| 日韩制服骚丝袜av| 中文字幕亚洲精品专区| 汤姆久久久久久久影院中文字幕| 免费日韩欧美在线观看| 女人久久www免费人成看片| 欧美中文综合在线视频| 一个人免费看片子| 菩萨蛮人人尽说江南好唐韦庄| 少妇人妻精品综合一区二区| 国产成人精品久久二区二区91 | 宅男免费午夜| 在线观看三级黄色| 久久人人97超碰香蕉20202| 久久天躁狠狠躁夜夜2o2o | 波多野结衣av一区二区av| 国产 一区精品| 色婷婷久久久亚洲欧美| 日韩熟女老妇一区二区性免费视频| 男女国产视频网站| 亚洲国产看品久久| 久久久久视频综合| 人人妻,人人澡人人爽秒播 | 一本色道久久久久久精品综合| 婷婷成人精品国产| 久久久久久久久久久久大奶| 欧美人与性动交α欧美软件| 亚洲av电影在线进入| 国产视频首页在线观看| 一区福利在线观看| 国产一区亚洲一区在线观看| 人人妻,人人澡人人爽秒播 | 国产精品免费大片| 巨乳人妻的诱惑在线观看| 午夜老司机福利片| 久久久久久久大尺度免费视频| 一级片免费观看大全| 人妻人人澡人人爽人人| 亚洲av电影在线进入| 精品一区二区三区av网在线观看 | 十分钟在线观看高清视频www| 老司机靠b影院| 色精品久久人妻99蜜桃| 丝袜脚勾引网站| 大陆偷拍与自拍| 色视频在线一区二区三区| 午夜福利乱码中文字幕| 国产亚洲欧美精品永久| 新久久久久国产一级毛片| 国产 一区精品| 久久久久国产精品人妻一区二区| 午夜福利网站1000一区二区三区| 日本欧美国产在线视频| 日韩熟女老妇一区二区性免费视频| 亚洲av电影在线观看一区二区三区| 男人添女人高潮全过程视频| 免费不卡黄色视频| 亚洲av成人不卡在线观看播放网 | 日韩人妻精品一区2区三区| 美女福利国产在线| 性高湖久久久久久久久免费观看| xxxhd国产人妻xxx| 国产成人一区二区在线| 看非洲黑人一级黄片| 97人妻天天添夜夜摸| 丝袜美腿诱惑在线| 9191精品国产免费久久| 国产免费现黄频在线看| 日日爽夜夜爽网站| 黄色一级大片看看| 国产av国产精品国产| 午夜福利,免费看| 国产精品三级大全| avwww免费| 成人国产av品久久久| 国产 一区精品| 亚洲精品成人av观看孕妇| 精品免费久久久久久久清纯 | 国产精品麻豆人妻色哟哟久久| 亚洲美女视频黄频| 亚洲av成人不卡在线观看播放网 | 亚洲人成电影观看| 老司机在亚洲福利影院| 国产福利在线免费观看视频| 老司机亚洲免费影院| 丝袜美腿诱惑在线| 亚洲欧美清纯卡通| 亚洲成色77777| 亚洲 欧美一区二区三区| 欧美老熟妇乱子伦牲交| 成人漫画全彩无遮挡| 日韩一区二区视频免费看| 色吧在线观看| 亚洲欧美色中文字幕在线| 日本wwww免费看| 亚洲成人一二三区av| 亚洲人成电影观看| 性色av一级| 母亲3免费完整高清在线观看| 又黄又粗又硬又大视频| 欧美日韩一区二区视频在线观看视频在线| 国语对白做爰xxxⅹ性视频网站| 90打野战视频偷拍视频| av国产久精品久网站免费入址| 日韩熟女老妇一区二区性免费视频| 免费观看性生交大片5| 久久精品aⅴ一区二区三区四区| 精品一区二区三区四区五区乱码 | 午夜影院在线不卡| 亚洲国产欧美在线一区| 日韩 欧美 亚洲 中文字幕| 国产精品亚洲av一区麻豆 | av网站在线播放免费| 97在线人人人人妻| 桃花免费在线播放| 国产成人系列免费观看| 精品午夜福利在线看| 亚洲av成人不卡在线观看播放网 | 超色免费av| 亚洲精品第二区| 亚洲一区二区三区欧美精品| 青春草国产在线视频| 亚洲av国产av综合av卡| 日韩中文字幕欧美一区二区 | 国产精品欧美亚洲77777| 精品一区二区三卡| 欧美国产精品一级二级三级| 宅男免费午夜| 久久久久久久精品精品| 美女高潮到喷水免费观看| 亚洲,一卡二卡三卡| 黄色视频不卡| 王馨瑶露胸无遮挡在线观看| 精品国产一区二区三区久久久樱花| 久久国产亚洲av麻豆专区| 久久鲁丝午夜福利片| 91老司机精品| 黄色一级大片看看| 精品国产国语对白av| 黑丝袜美女国产一区| 久久精品国产综合久久久| 精品国产一区二区三区久久久樱花| 久久精品国产综合久久久| 色网站视频免费| 日韩欧美精品免费久久| 亚洲欧美一区二区三区黑人| 国产精品久久久久久精品电影小说| 欧美在线一区亚洲| 精品少妇一区二区三区视频日本电影 | 国产亚洲欧美精品永久| 国产av一区二区精品久久| av一本久久久久| 18禁国产床啪视频网站| 最近中文字幕高清免费大全6| 久久精品国产亚洲av涩爱| 啦啦啦中文免费视频观看日本| 不卡视频在线观看欧美| 黄色毛片三级朝国网站| 国产精品av久久久久免费| av国产久精品久网站免费入址| 最近中文字幕2019免费版| 亚洲欧美清纯卡通| 男的添女的下面高潮视频| 亚洲欧美精品综合一区二区三区| 亚洲人成电影观看| 18禁动态无遮挡网站| bbb黄色大片| 人人妻,人人澡人人爽秒播 | 丰满迷人的少妇在线观看| 久久亚洲国产成人精品v| 精品第一国产精品| 精品酒店卫生间| 热re99久久国产66热| 99精国产麻豆久久婷婷| 久久精品国产综合久久久| 在线天堂中文资源库| 久久热在线av| 大香蕉久久网| 无限看片的www在线观看| 高清黄色对白视频在线免费看| 欧美日韩精品网址| 亚洲美女黄色视频免费看| 久久久久久人人人人人| 七月丁香在线播放| 午夜福利乱码中文字幕| 久久久久久久久免费视频了| 亚洲av日韩精品久久久久久密 | 最近的中文字幕免费完整| 日韩大码丰满熟妇| 久久久久久人人人人人| 十分钟在线观看高清视频www| 国产又爽黄色视频| 亚洲精品国产色婷婷电影| 男女床上黄色一级片免费看| 男男h啪啪无遮挡| 老司机影院毛片| 老司机深夜福利视频在线观看 | av.在线天堂| 曰老女人黄片| 大香蕉久久成人网| 亚洲专区中文字幕在线 | 蜜桃在线观看..| 人人澡人人妻人| 黄色视频不卡| 我要看黄色一级片免费的| 精品福利永久在线观看| 麻豆精品久久久久久蜜桃| 美女主播在线视频| 无遮挡黄片免费观看| 2018国产大陆天天弄谢| 欧美变态另类bdsm刘玥| 久久97久久精品| 黑丝袜美女国产一区| 五月天丁香电影| 1024视频免费在线观看| 中文字幕亚洲精品专区| 爱豆传媒免费全集在线观看| 赤兔流量卡办理| 久久99热这里只频精品6学生| 久久久久久久久久久免费av| av不卡在线播放| 亚洲精品视频女| 老司机靠b影院| 色婷婷av一区二区三区视频| 秋霞伦理黄片| 亚洲精品国产av成人精品| 国产av精品麻豆| 国产精品嫩草影院av在线观看| 日韩人妻精品一区2区三区| 免费黄色在线免费观看| 69精品国产乱码久久久| 国产精品麻豆人妻色哟哟久久| 最近最新中文字幕免费大全7| 18禁动态无遮挡网站| 日本色播在线视频| 亚洲国产精品成人久久小说| 国产成人精品在线电影| 校园人妻丝袜中文字幕| 精品国产一区二区三区四区第35| 色吧在线观看| 午夜福利视频在线观看免费| 1024视频免费在线观看| av不卡在线播放| 日韩熟女老妇一区二区性免费视频| 国产成人免费观看mmmm| 国产精品久久久av美女十八| 狂野欧美激情性xxxx| 如何舔出高潮| 精品国产乱码久久久久久小说| 超碰97精品在线观看| 午夜av观看不卡| 婷婷色av中文字幕| 亚洲av国产av综合av卡| 建设人人有责人人尽责人人享有的| 麻豆av在线久日| 啦啦啦啦在线视频资源| 欧美日韩一级在线毛片| 欧美亚洲日本最大视频资源| 国产黄色免费在线视频| 国产片特级美女逼逼视频| 中文天堂在线官网| 午夜激情av网站| 精品福利永久在线观看| 色婷婷久久久亚洲欧美| 精品国产一区二区三区久久久樱花| 好男人视频免费观看在线| 99国产综合亚洲精品| 在现免费观看毛片| 亚洲欧美精品自产自拍| 天天躁日日躁夜夜躁夜夜| 精品酒店卫生间| 亚洲色图综合在线观看| 深夜精品福利| 伦理电影免费视频| 久久天堂一区二区三区四区| 久久久精品94久久精品| 1024香蕉在线观看| 伦理电影免费视频| 精品视频人人做人人爽| 国产亚洲午夜精品一区二区久久| 欧美变态另类bdsm刘玥| 久久综合国产亚洲精品| 国产又色又爽无遮挡免| 天堂8中文在线网| 人人妻人人澡人人爽人人夜夜| 免费在线观看视频国产中文字幕亚洲 | 2021少妇久久久久久久久久久| 国产伦人伦偷精品视频| 一本久久精品| 免费女性裸体啪啪无遮挡网站| 99久国产av精品国产电影| 在线看a的网站| 老司机在亚洲福利影院| 久久免费观看电影| 在线免费观看不下载黄p国产| 最新的欧美精品一区二区| 国产精品国产三级国产专区5o| 亚洲一区二区三区欧美精品| 一边摸一边做爽爽视频免费| 婷婷色麻豆天堂久久| 国产精品国产三级国产专区5o| 伊人亚洲综合成人网| 一边摸一边抽搐一进一出视频| 天堂中文最新版在线下载| 一边摸一边抽搐一进一出视频| 母亲3免费完整高清在线观看| 亚洲精品第二区| 香蕉国产在线看| 在线亚洲精品国产二区图片欧美| 午夜老司机福利片| 亚洲国产毛片av蜜桃av| 多毛熟女@视频| 人人妻人人爽人人添夜夜欢视频| 久久精品久久精品一区二区三区| 男女免费视频国产| 如何舔出高潮| 国产成人啪精品午夜网站| 黄色一级大片看看| 最近2019中文字幕mv第一页| 日本wwww免费看| 亚洲精华国产精华液的使用体验| 丝袜美足系列| 久久99精品国语久久久| 无遮挡黄片免费观看| 日韩成人av中文字幕在线观看| 亚洲av日韩在线播放| 国产欧美日韩一区二区三区在线| 国产麻豆69| 亚洲第一av免费看| svipshipincom国产片| 成年人免费黄色播放视频| 免费在线观看视频国产中文字幕亚洲 | 免费在线观看视频国产中文字幕亚洲 | 大片免费播放器 马上看| 亚洲一码二码三码区别大吗| 国产亚洲av高清不卡| 亚洲av成人精品一二三区| 久久国产精品男人的天堂亚洲| 黑人欧美特级aaaaaa片| 国产男女超爽视频在线观看| 五月天丁香电影| 久久天躁狠狠躁夜夜2o2o | 高清在线视频一区二区三区| 十分钟在线观看高清视频www| 欧美97在线视频| 中文字幕最新亚洲高清| 亚洲国产日韩一区二区| netflix在线观看网站| 亚洲av日韩在线播放| 精品少妇一区二区三区视频日本电影 | bbb黄色大片| 亚洲欧美色中文字幕在线| 亚洲精品国产色婷婷电影| 大香蕉久久网| 成年人午夜在线观看视频| 日韩大码丰满熟妇| 亚洲国产成人一精品久久久| 国产精品秋霞免费鲁丝片| 免费黄网站久久成人精品| 国产极品粉嫩免费观看在线| 欧美av亚洲av综合av国产av | 亚洲,欧美精品.| 丝袜美足系列| 欧美精品一区二区大全| 天堂俺去俺来也www色官网| 精品久久久久久电影网| 精品人妻在线不人妻| 大码成人一级视频| 中文字幕精品免费在线观看视频| 欧美精品一区二区大全| 亚洲av综合色区一区| 日韩不卡一区二区三区视频在线| 伊人久久国产一区二区| 最新在线观看一区二区三区 | 一级毛片黄色毛片免费观看视频| 咕卡用的链子| 男的添女的下面高潮视频| 韩国高清视频一区二区三区| 少妇人妻 视频| 国产色婷婷99| 久久人人爽av亚洲精品天堂| 老司机影院毛片| 狂野欧美激情性bbbbbb|