• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep reinforcement learning based task offloading in blockchain enabled smart city①

    2023-09-12 07:30:00JINKaiqi金凱琦WUWenjunGAOYangYINYufenSIPengbo
    High Technology Letters 2023年3期

    JIN Kaiqi (金凱琦),WU Wenjun,GAO Yang,YIN Yufen,SI Pengbo

    (Faculty of Information Technology,Beijing University of Technology,Beijing 100124,P.R.China)

    Abstract With the expansion of cities and emerging complicated application,smart city has become an intelligent management mechanism.In order to guarantee the information security and quality of service(QoS) of the Internet of Thing(IoT) devices in the smart city,a mobile edge computing (MEC) enabled blockchain system is considered as the smart city scenario where the offloading process of computing tasks is a key aspect infecting the system performance in terms of service profit and latency.The task offloading process is formulated as a Markov decision process (MDP) and the optimal goal is the cumulative profit for the offloading nodes considering task profit and service latency cost,under the restriction of system timeout as well as processing resource.Then,a policy gradient based task offloading (PG-TO) algorithm is proposed to solve the optimization problem.Finally,the numerical result shows that the proposed PG-TO has better performance than the comparison algorithm,and the system performance as well as QoS is analyzed respectively.The testing result indicates that the proposed method has good generalization.

    Key words: mobile edge computing(MEC), blockchain, policy gradient, task offloading

    0 Introduction

    Developed with the help of big data and Internet of Thing (IoT)[1],the smart city is capable of the intelligence management in terms of smart grid,smart community, smart hospitality, smart transportation,smart warehouse and smart healthcare[2].However,most of the IoT devices in smart city are lack of integrated security mechanisms and vulnerably exposed in the open area,which brings challenges such as data privacy and security into smart city[3-4].

    Fortunately,blockchain,which is an intelligent decentralized system using distributed databases to identify,disseminate and record information,makes it possible to guarantee the data security and establish a reliable network system[5].The core techniques of blockchain are consensus mechanism and smart contract,which are responsible for the development of trust among distributed nodes and autonomic management of the system,respectively.Due to the intervention of blockchain,the security access[6],data privacy and security[7-8]and data integrity[9]can be ensured in smart city.

    However,it is wildly acknowledged that the operation of consensus mechanism and smart contract of blockchain requires substantial computation resource.To further facilitate the implementation of blockchain in smart city,mobile edge computing (MEC) servers are always deployed with the access points to provide highly expandable computation resource[10]and reduce the service latency[11].

    In the MEC-enhanced smart city,the wireless resource allocation problem has been wildly studied.The dynamic scheduling problem is researched in the Internet of Everything,and a multiple scheduling algorithm is proposed adopting round robin scheduling,proportional fair scheduling and priority based scheduling,enhancing the resource allocation efficiency[12].The joint user association and data rate allocation problem is modeled to optimize the service delay and power consumption,and correspondingly solved by iterative algorithm and bisection algorithm[13].Task offloading is the most typical resource management problem in MEC,and it is also widely studied in the scenario of smart city.In a multi-user and multi-server MEC scenario,a joint task offloading and scheduling problem is studied aiming to minimize task execution latency and solved by a heuristic algorithm[14].Aiming at the maximization of offloading accomplish rate,problems restricted by the power consumption of mobile equipment as well as the deadline of offloading tasks are studied by adopting deep reinforcement learning (DRL) and convex optimization[15-16].Considering the time cost,energy cost of IoT devices and resource utilization of cloudlets in smart city, the computation offload problem is researched,and a balanced offload strategy is perceived using Pareto optimization[17].By using game theory,the comprehensive profit of mobile devices and edge service providers are designed as optimization problems,coming up with an offloading solution enhancing both data privacy and system power efficiency[18].

    When blockchain is enabled in smart city,the resource management problem is necessarily considered with more variables such as task profit obtained from the blockchain system,service latency and resource occupation.To optimize these multiple optimization objectives with complex restrictions,DRL is efficient by the perception ability of deep learning and the decisionmaking ability of reinforcement learning[19-20].In blockchain enabled MEC scenario,task scheduling and offloading problems are studied and modified as Markov decision process(MDP).Policy gradient method is adopted to solve the optimization goal considering the long-term mining reward,system resource and latency cost while deep Q-learning (DQL) optimizes the task execution delay and energy consumption[21-22].Another DRL method called asynchronous advantage actor-critic(A3C) is applied in the cooperative computation offloading and resource allocation scenario,enhancing the computation rate and throughput as main factors of optimization function[23].In the intelligent resource allocation mechanism for video services,the system performance is optimized by A3C in terms of throughput and latency[24].Aiming at the system energy efficiency in Internet of Vehicles (IoV) scenario,A3C is also adopted to solve the optimization function composed of energy and the computation overheads[25].

    Specifically,in a related scenario proposed in Ref.[26],where MEC provides users with sufficient resources to reduce the computing pressure in the IoT system supported by blockchain,the node matching between users and edge computing servers in consideration of wireless channel quality and QoS is studied,the simulation results show excellent performance with the help of reinforcement with baseline algorithm.However,different from the common IoT,the smart city has diverse requirements in application layer,and the demand of high restrictions on latency in applications such as high definition (HD) mapping and intelligent driving decision in smart traffic service is one of the most important factors[27].In this perspective,the application implied in the smart city dabbles in a wider range and ensures more latency restriction in comparison with common IoT services,which has not been fully studied.

    In order to realize the data security and ensure the quality of service(QoS)requirements of smart city applications,both the blockchain and MEC technologies are adopted to enhance the system.The secure process and storage of original data from the smart city applications with latency constraint is considered as the task.The task offloading problem maximizing the long-term reward which consists of the profits of task processing and the cost of task processing latency is the main concern.The task offloading problem is formulated as a MDP,and the DRL method which uses a deep neural network as the decision agent is used to optimize the long-term reward.The episodic simulation is built,and the policy gradient (PG) with baseline algorithm is adopted to train the decision agent.The performance with different environment parameters are tested,which respectively confirm the effectiveness and the generalization ability of the policy gradient based task offloading(PG-TO)algorithm.

    The rest of this article is organized as follows.System model is formed in Section 1.The task offloading problem is formulated in Section 2 with the definition of the actions,states and rewards of the MDP.In Section 3,DRL is used to solve the task offloading problem.The training and testing results are given in Section 4.At last,this paper is concluded in Section 5.

    1 System model

    In this paper,the MEC-enhanced smart city scenario is considered,and the blockchain technology is adopted to ensure the secure process and storage of original data from the smart city applications.As shown in Fig.1,the system is composed of physical layer and blockchain layer,which is described in detail in this section.Besides,the reputation model and the profit model are also given in the following.

    1.1 Physical layer

    The physical layer of the proposed MEC-enhanced smart city with blockchain management functions is shown in Fig.1.There are 3 classes of participants in the proposed physical layer,which are user equipment(UE),access point (AP),and main access point(MAP).

    (1) UE can be IoT device or mobile device which has limited frequency and computing resources.It is the user of various emerging applications in smart city,and the key information of its applications or behaviors are considered as the computation task which needs to be uploaded to the blockchain for secure process and storage.It is capable of operating the preset program according to the smart contract in blockchain but does not participate in the consensus process.

    (2) AP is generally the small base station deployed with MEC server.It performs as the blockchain node which takes part in the consensus process in blockchain and is responsible for uploading the offloaded tasks from UEs to the blockchain.

    (3) MAP is the central control node which is in charge of assigning the computation tasks form UEs to appropriate APs.Besides,it has the similar property as AP that it can participate in the consensus process and maintain the blockchain system.

    1.2 Blockchain layer

    As the rules of blockchain nodes running in the smart city,the smart contract defines the detailed responsibility of the nodes.Based on the smart contract,the specific steps of the task offloading process are shown below.

    (1) UE adopts various applications in smart city generating computation task and declares the average task fee it can pay for the task.

    (2) UE generates task offloading request including information of the average task fee and the latency limitation,and sends the requests to the MAP.

    (3) MAP observes the frequency and computing resource occupation of all the APs,the reputation value of the APs and the arrived task offloading requests of the UEs.

    (4) MAP assigns the tasks to APs according to the current system observation in step(3).

    (5) APs allocate frequency and computing resources to transmit and compute the offloaded tasks.

    (6) If the task is offloaded and computed successfully within the tolerable latency,the task is accomplished and task fee is paid to the APs as task profit.Otherwise,AP does not get paid.

    (7) MAP updates the reputation value of APs according to the accomplish status of tasks.

    1.3 Reputation model

    In order to value the reputation and profit for each APs,the corresponding value model is proposed in the following subsections.As the historical reputation of each blockchain node reflects its competitiveness in the consensus process,the credit model formulated in a MEC enabled blockchain system[26]is used in this paper.The reputation is denoted byri,which shows the reliability of thei-th AP and is closely related to the probability of accomplishing a task within the tolerable latency.The value ofriis linear within a restricted rangeri∈[1,5] ,whose initial value is set as 3.Every time if the task is accomplished by thei-th AP,(ri+0.1),otherwise,(ri-0.5).

    1.4 Profit layer

    The profit of APs is only contributed by the task fee paid from the offloaded tasks of UEs.In order to establish an incentive mechanism,it is supposed that APs that have high reputation value will be paid with more profit,which is a reward and approval for the high quality of service.When thei-th AP accomplishes thej-th offload service,the practical profit can be described as

    2 Problem formulation

    In the blockchain enabled smart city scenario,the frequency and computing resources of the APs are limited.As the task offloading is a key process which can affect not only the profit of the APs but also the overall system performance,it is formulated as an optimization problem.The target is set to be maximizing the longterm cumulative profit gains from offloaded tasks for APs while considering the limitation of the resources of APs and the service latency constraint for QoS.The optimization problem is formulated as

    According to the system model described in Section 1,as MAP only observes the situation of current time step,the task offloading process has the Markov property, which is modeled as a Markov decision process (MDP).The MDP is denoted as(S,A,P,R) ,whereSstands for state space,andArepresents the action space.Specifically,Pis defined asP(s′|s,a) to describe the probability of the system state transition froms∈Stos′∈Safter taking actiona∈A,whileRis defined asR(a,s) which is the reward from taking actionawhen state iss.The detailed definition of state,action and reward is shown below.

    2.1 State

    The design of state should consider those elements includingSA,SR,SOandSB.SArepresents the state of the APs,which includes the frequency and computing resource occupation situation of the current time and the next(T-1) time steps.SRdenotes the state of the processing tasks in the system,including the index and reputation value of the AP where the tasks are offloaded,and the practical profit gain for AP when tasks are accomplished.SOandSBgive the state of tasks which have arrived and are waiting to be offloaded.SOdenotes the tasks which can be observed in detail,including information of the required frequency resource in each APs if offloaded to,the related transmission latency and the expected profit.AndSBdenotes the extra unobservable tasks that have arrived at the system,including the arrival time,expected profit and the expected service latency of these tasks.The state structure detail is shown in Fig.2.

    Fig.2 The state observation of each time step t

    2.2 Action

    In the researched scenario,as MAP is capable of observing state information of APs,the definition of action is to select the suitable APs for UEs to offload their tasks.In time stept,if the await offloaded tasks number isDtand the APs amount isNA,the action space size is (NA+1)Dt,for the reason that the task could be either offloaded to any of the APs or not.Therefore,the action space is so huge that the efficiency of the proposed algorithm would decrease.As a result,simplify the action space by observing the firstKnumber of await tasks at each time step and offload them adopting first-in-first-out (FIFO) method.

    As the intellectual property of task offloading,the optimization of long-term accumulation value by serial decision could be realized the same as the parallel decision.Hence,it is considered that the parallel action decisions are decomposed into serial decisions in the same time step.The decision times in time steptis denoted asXtand all the action in the time steptis denoted as

    In this method could the action space be reduced,whose space size isNA+1 ,denoting as{1,2,…,NA,?} ,whereat,x=imeans that the first await task will be processed by thei-th AP in thex-th step at timet.Specifically,if the first task is thej-th task in the system task process list,at,x=istands forat,ji= 1.

    2.3 Reward

    As the optimization problem is defined in Eq.(3),the computing and frequency resources are easily implemented that MAP will not offload tasks to the APs whose frequency and computing resources are insufficient.As for the service experience,the service latency restriction of tasks in Eq.(3) can be defined as a penalty function in the reward to guarantee the accomplish rate of the scheme.Therefore,the reward can be defined based on Eq.(3) as

    where the discount factor isδ= 1 ,andTMdenotes the maximum time step number of one episodic simulation.

    3 Policy gradient based task offloading algorithm

    This section introduces the PG with baseline which is one of the DRL algorithms,to develop the policy gradient-based task offloading algorithm (PG-TO)for the proposed optimization problem.The PG-TO algorithm is illustrated in detail including the adoption of the PG with baseline method and the episodic simulation process in the following.

    3.1 The adoption of PG with baseline

    The adoption of PG with baseline algorithm in the proposed optimization problem and the data flow of state,action and reward in the training process are shown in Fig.3.The PG agent,which is the policy networkπθ,outputs the probability distribution of action according to the input state,and then randomly chooses the action based on the probability distribution.The network parameterθinπθis continuously optimized after each training iteration.In this way,the probability distribution of actions under each state can be established and approach the optimal strategy of probability distribution of action.

    As the average profit of each time step for theq-th sample,the baseline is calculated for eachEtimes of episodic simulation in one iteration.The variance is reduced and the efficiency of policy training is enhanced by the deduction of baseline fromvi,l.

    Fig.3 The data flow in the training of the proposed PG-TO algorithm

    3.2 Simulation design

    The overall episodic simulation design of the proposed PG-TO algorithm is shown as Algorithm 1.The total iteration number is set asPand the policy networkπθis trained based on the PG with baseline.

    Algorithm 1 PG-TO 1: Establish UEs,APs and MAP in the simulation environment,initialize the related system parameters.2: Generate Q worksheets.3: Randomly initialize the neural network parameters θ by normal distribution.4: for iteration p < P do 5: for episodic simulation i < I do 6: get the initial state si,l = si,0 7: while time step t < TM do 8: based on πθ and si,l get ai,l 9: if action ai,j =?then calculate kt according to Eq.(5)update t = t +1 10: else Offload the j-th task to the ai,l -th APs kt = 0 acquire f^j ,t^f,j 11: end if 12: store si,j,ai,j,ri,j in the trajectory 13: update J1(t) ,J2(t)14: update state si,l 15: end while 16: end for 17: Calculate the baseline b「i/E?,l by Eq.(8)18: Update parameters of network by Eq.(7)19: end for

    In the beginning of each time of episodic simulation,the policy networkπθtakes observation of current simulation environment gettingsi,land output the probability distribution of action.Then, actionai,lis selected randomly according to the output probability distribution of action.Ifai,l=?, there is no task offloaded to any APs at current time step, is obtained according to Eq.(5) and then the system time step moves on.If the related task is offloaded to theai,l-th AP for processing, whilef^jandt^f,jare counted, andri,l=0.Then,{si,l,ai,l,ri,l}is recorded as one sample in the trajectory.Afterwards, the related task setsare updated and the next statesi,l+1is obtained.The above steps are repeated to obtain a complete trajectory sample of a whole episodic simulation as long ast

    4 Simulation and performance evaluation

    4.1 Training performance

    The training parameters are defined in Table1.Refer to the scenario setting in Ref.[26],which adopts NB-IoT service model specified in 3GPP 36.752[28].The bandwidth is set as 180 kHz,and the number of frequency resource units is 48 as 3.75 kHz single-tone is adopted.The size of task and the corresponding profit for APs are classified into small and big referring to Ref.[26].The transmission frequency resource requirements of tasks are also defined according to the subchannel classifications in NB-IoT.The observable future timeTis set as 300 times steps as the system performance indicators are more reasonable when the system reaches a steady state and to avoid excessive redundant simulation experiments.The number of APsNAin the simulation makes trade-off between system state complexity and reasonable service pressure.The numerical setting ofNTmakes a reasonable size of input state space while the number ofNWcomprehensively considers the APs' numbers,overall system tasks quantity and job density.

    Symbol Parameter Setting D Maximum frequency resource units number 48 F Maximum computing resource units number 64(Dt,Ft)Frequency resource requirement and transmitting time of j-th data package in each time step (big){(4,8),(1,8),(48,1)}{(24,2),(12,4),(4,2)}T Time steps that can be observed to the future 300 Frequency resource requirement and transmitting time of j-th data package in each time step (small)fj Expected task profit of the j-th task(big) [10,15]Expected task profit of the j-th task(small) [15,20]T~ Maximum latency of tasks 25 NA Number of APs 5 NT Maximum number of observable processing tasks of each AP 25 NW Maximum storage number for await allocate tasks arrived of each AP 125 α Weight for task profit 1 β Latency cost weight for each task 5 λ Tasks arrival rate 0.8 R Proportion of small tasks in all tasks 0.8

    The policy network used by PG-TO method has 2 fully hidden layers,each of which has 32 neurons.Moreover,max-SINR,max-credit,max-resource and random are set as the comparison methods.Max-SINR always selects the AP which has the maximum SINR towards UEs.The max-credit prefers the AP with the maximum reputation valueriwhile the max-resource tends to offload tasks to APs with the maximum available frequency and computing resources.

    Fig.4 Mean reward φ

    Fig.4 shows the average value of mean rewardφwhich is the averageg(a) ofIsamples in each iteration.For training efficiency,the learning rate is set as 0.001.And then,it is decreased to 0.0005 after 20-th iteration for better convergence performance.Theφof PG-TO rises rapidly and surpasses the value of the max-credit after only 5 iterations,and finally converges to about 9100.Theφof PG-TO is about 16.6% better than the second-best max-credit strategy,while max-SINR strategy is in the third place,about 7180.The random strategy and the max-resource strategy perform worst,both of which range from 6500 to 6800.

    As two important components inφ,the average task profit and the average latency cost of APs are shown in Fig.5,and their calculation formula is shown in Table 2.In Fig.5(a),the max-credit strategy obtains the largest income among all the comparison algorithms as it always picks the AP with the highest reputation and earns higher task profit for each completed task according to Eq.(1).Meanwhile,max-SINR obtains the lowest latency cost due to the good quality of wireless transmission referring to Fig.5(b).In Fig.5 and Fig.4,it can be found that there is a contradiction between the profit of APs and the latency performance of UEs.However,the advantage in task profit surpasses the shortage of latency cost,so the max-credit reaches the second-best strategy.Specifically for PG-TO,the optimal performance inφis contributed by a little bit better task profit and about 40% less latency cost compared with max-credit.

    Indicators Calculation formula Figure Task profit α ∑f^j∈Jd(t)j(at)t^f,j + td,j[] Fig.5(a)Latency cost β ∑1 j∈Ja(t)[] Fig.5(b)tf,j + td,j Overtime rate ∑j∈Ja(t)(t^f,j + td,j)/(tf,j + td,j) Fig.6(c)Overfee rate ∑j∈Jd(t)f^j/fj Fig.7

    Moreover,the indicators which closely related to the system performance are also calculated and analyzed in Fig.6,and the related calculation formula of Fig.6 (c) is shown in Table.2.As shown in Fig.6(a),the frequency resource occupation of PG-TO is on the same level as the comparison strategies except for max-SINR.This is because the max-SINR strategy causes load imbalance among APs,and thus restricts the utilization rate of frequency resources.

    As shown in Fig.6(b) and Fig.6(d),the performances of computing resource occupation and the complete rate are consistent.Specifically,in Fig.6(d),the mean complete rate is calculated in the simulation program,which is the ratio of the number of finished tasks to the total number of tasks in the system.It can be observed that the max-credit strategy completes the fewest tasks while the other comparison algorithms complete more amount of tasks and consumes more amount of computing resources.

    Besides,the ratio of practical task profit to expected task profit is further counted as the overfee rate.Accordingly,the overfee rate range of each strategy is shown in Fig.7,which marks the corresponding maximum and minimum overfee rate of each iteration and the related calculation formula is shown in Table.2.Generally,the mean overfee rate of PG-TO algorithm fluctuates in the range of 1.16 to 1.27,and its mean overfee rate is 14.2% higher than the second-best max-credit strategy,lower than any other comparison strategies.Moreover,it is easily spotted that the variance overfee rate of max-credit and that of PG-TO are at a significantly high level compared with other strategies.The reason is that max-credit strategy always selects the AP with the maximum reputation value,which causes tasks backlogs in the high credit value APs whose reputation declines due to the timeout of tasks.However,as shown in Fig.5(a) and Fig 6(d),maxcredit strategy has greater advantages than other comparison strategies with the lowest complete rate.Hence,it could be indicated that tasks with relatively high initial task fee are opportunistically offloaded to the higher reputation APs while tasks with lower initial task fee are selectively abandoned.Similarly,the intelligent task offloading scheme learned by PG-TO is shown by the training results that it can gradually learn the advantages of the max-credit and alleviate the huge cost of latency as shown in Fig.5(b).

    Fig.7 Overfee rate range

    4.2 Test performance

    For the policy network trained under the situation that the proportion of small tasks is 0.8,the performance with proportion of small tasks chaning is tested.In the test of each small tasks proportion,100 samples different from the training samples are generated for simulation.

    Fig.8 shows the test results in the proportion of small tasks from 72% to 88% with 2% interval.As the small tasks proportion increases,the total task profit decreases,thenφvalues of random strategy,max-resource strategy,max-SINR strategy and PG-TO all show a downward trend as a whole accordingly except for the max-credit strategy.

    For the max-credit strategy,φis positively related to the increase proportion of the small tasks.That is because the overtime problem caused by the unbalanced offloading scheme can be slightly alleviated when the overall load of the system reduces,and then the advantage of choosing the maximum reputation to get the maximum practical task profit becomes obvious.Additionally,PG-TO obtains the highestφwhen the small tasks proportion ranges from 0.72 to 0.86,which indicates a good generalization.

    Fig.8 Test result of mean reward φ

    In all,the proposed PG-TO algorithm has a significant advantage in the optimization goal over the other 4 comparison algorithms.The PG-TO algorithm is capable of providing APs with more task profit while maintaining an acceptable latency cost.The resource occupation of PG-TO is at an average value,and the overtime rate along with complete rate are well balanced.Moreover,PG-TO shows a good generalization facing with different level of system tasks load.This means that the proposed PG-TO algorithm is capable of intelligently selecting proper APs to have the tasks offloaded in comprehensive consideration of APs' state,wireless environment,and tasks set while ensuring the latency requirements.

    5 Conclusion

    This article researches the task offloading process in a MEC-enhanced smart city with blockchain management functions.The task offloading process is modeled as a MDP and an optimization problem is developed focusing on the profit gain for APs and QoS requirement of UEs.The proposed optimization problem is solved by the PG method using the reinforce with baseline algorithm, and its training performance is 16.7% better than the second-best comparison strategy.The test performance with various small task proportion indicates that the proposed PG-TO algorithm has a good generalization.

    在线观看66精品国产| 午夜福利在线在线| 久久天堂一区二区三区四区| cao死你这个sao货| 免费看日本二区| 脱女人内裤的视频| 又粗又爽又猛毛片免费看| 欧美黄色片欧美黄色片| tocl精华| 最近最新中文字幕大全电影3| 国产精品乱码一区二三区的特点| 伊人久久大香线蕉亚洲五| 男女床上黄色一级片免费看| 成年女人永久免费观看视频| 久久久国产精品麻豆| 老鸭窝网址在线观看| 首页视频小说图片口味搜索| 人人妻人人看人人澡| 我的老师免费观看完整版| 色吧在线观看| 好男人在线观看高清免费视频| 伦理电影免费视频| 又紧又爽又黄一区二区| 亚洲,欧美精品.| 午夜福利在线观看吧| 成人亚洲精品av一区二区| 久久精品亚洲精品国产色婷小说| 啦啦啦免费观看视频1| 午夜亚洲福利在线播放| 精品久久久久久成人av| 亚洲精品一卡2卡三卡4卡5卡| 啦啦啦免费观看视频1| or卡值多少钱| 午夜激情欧美在线| 国产精品电影一区二区三区| 好男人电影高清在线观看| 国产欧美日韩精品亚洲av| 黄片大片在线免费观看| 看免费av毛片| 此物有八面人人有两片| 久久久精品欧美日韩精品| 99热6这里只有精品| 法律面前人人平等表现在哪些方面| 国产伦一二天堂av在线观看| 色在线成人网| 亚洲欧美一区二区三区黑人| 亚洲精品乱码久久久v下载方式 | 亚洲国产看品久久| 亚洲精品久久国产高清桃花| 精品久久久久久久久久免费视频| 夜夜看夜夜爽夜夜摸| 丰满的人妻完整版| 午夜福利在线观看吧| 国产av一区在线观看免费| 成年版毛片免费区| 免费电影在线观看免费观看| 麻豆成人午夜福利视频| 亚洲国产日韩欧美精品在线观看 | 九九热线精品视视频播放| 热99re8久久精品国产| 一级毛片高清免费大全| 村上凉子中文字幕在线| 久久中文字幕人妻熟女| 亚洲av电影不卡..在线观看| 国产亚洲av高清不卡| 国产精品久久久久久精品电影| a级毛片在线看网站| 亚洲国产中文字幕在线视频| 国产极品精品免费视频能看的| 97超视频在线观看视频| 亚洲人与动物交配视频| 亚洲va日本ⅴa欧美va伊人久久| 精品久久久久久,| 在线观看美女被高潮喷水网站 | 岛国在线免费视频观看| 一卡2卡三卡四卡精品乱码亚洲| 757午夜福利合集在线观看| 啦啦啦免费观看视频1| 免费人成视频x8x8入口观看| 在线播放国产精品三级| ponron亚洲| 亚洲片人在线观看| 天天添夜夜摸| 黑人巨大精品欧美一区二区mp4| 国产精品99久久久久久久久| 亚洲人成网站在线播放欧美日韩| 麻豆久久精品国产亚洲av| 亚洲美女黄片视频| 久久精品国产清高在天天线| 俄罗斯特黄特色一大片| 男插女下体视频免费在线播放| 中出人妻视频一区二区| 成人一区二区视频在线观看| 欧美不卡视频在线免费观看| 亚洲 国产 在线| 99久久精品热视频| 久久久精品欧美日韩精品| 天天躁日日操中文字幕| 欧美色欧美亚洲另类二区| 午夜精品一区二区三区免费看| 黄色 视频免费看| 女生性感内裤真人,穿戴方法视频| 精品熟女少妇八av免费久了| 高潮久久久久久久久久久不卡| 久久久久久人人人人人| 制服丝袜大香蕉在线| 99久国产av精品| 欧美日韩精品网址| 床上黄色一级片| 精品一区二区三区四区五区乱码| 色老头精品视频在线观看| 最新中文字幕久久久久 | 观看免费一级毛片| 久久久久亚洲av毛片大全| 少妇裸体淫交视频免费看高清| 久久久久国产一级毛片高清牌| 亚洲午夜精品一区,二区,三区| 在线永久观看黄色视频| 麻豆国产av国片精品| 亚洲av中文字字幕乱码综合| 91av网一区二区| 亚洲五月天丁香| 观看免费一级毛片| 村上凉子中文字幕在线| 美女cb高潮喷水在线观看 | 亚洲精品中文字幕一二三四区| 欧美极品一区二区三区四区| 欧美日韩黄片免| 欧美日韩中文字幕国产精品一区二区三区| 亚洲av电影不卡..在线观看| 国产精品综合久久久久久久免费| 狠狠狠狠99中文字幕| 亚洲国产精品久久男人天堂| 日韩精品青青久久久久久| 少妇裸体淫交视频免费看高清| 国产成人啪精品午夜网站| 久久香蕉精品热| 天堂√8在线中文| 久久久久性生活片| 国产精品国产高清国产av| 国产精品久久久人人做人人爽| 国产亚洲精品久久久com| 啦啦啦观看免费观看视频高清| 嫩草影视91久久| 日韩欧美免费精品| 国产aⅴ精品一区二区三区波| 欧美一区二区精品小视频在线| 露出奶头的视频| 色综合欧美亚洲国产小说| 亚洲色图av天堂| 在线免费观看的www视频| 国产美女午夜福利| 麻豆成人av在线观看| 亚洲av成人av| 成熟少妇高潮喷水视频| 一进一出抽搐gif免费好疼| 精品久久久久久久毛片微露脸| 婷婷丁香在线五月| 国内少妇人妻偷人精品xxx网站 | 成人av一区二区三区在线看| 国产成人精品无人区| 少妇的丰满在线观看| 午夜成年电影在线免费观看| 国产高清有码在线观看视频| 亚洲精华国产精华精| 亚洲成av人片免费观看| 国产69精品久久久久777片 | 日本黄大片高清| 男女下面进入的视频免费午夜| 国产精品精品国产色婷婷| 国产免费av片在线观看野外av| 精品国产乱码久久久久久男人| 一进一出抽搐动态| 黑人欧美特级aaaaaa片| 亚洲精品一卡2卡三卡4卡5卡| 亚洲国产看品久久| 欧美在线一区亚洲| 又爽又黄无遮挡网站| 国产av麻豆久久久久久久| 午夜福利高清视频| 亚洲五月天丁香| 国产1区2区3区精品| 国产精品久久久av美女十八| 麻豆国产av国片精品| 欧美绝顶高潮抽搐喷水| 国产av不卡久久| 亚洲精品456在线播放app | 国产淫片久久久久久久久 | 欧美激情久久久久久爽电影| 久久国产乱子伦精品免费另类| 99国产精品一区二区蜜桃av| 中文字幕最新亚洲高清| 亚洲一区二区三区色噜噜| 亚洲熟妇熟女久久| 国产成人影院久久av| www.熟女人妻精品国产| 国内精品美女久久久久久| 国产欧美日韩精品亚洲av| 99久久国产精品久久久| 久久精品国产综合久久久| 两个人的视频大全免费| 88av欧美| 国产成人一区二区三区免费视频网站| 国产伦在线观看视频一区| 亚洲国产精品合色在线| 黄色成人免费大全| 宅男免费午夜| 两性夫妻黄色片| 国内久久婷婷六月综合欲色啪| 国产午夜福利久久久久久| 在线永久观看黄色视频| 免费人成视频x8x8入口观看| 日本撒尿小便嘘嘘汇集6| 亚洲在线观看片| 999久久久精品免费观看国产| 美女高潮喷水抽搐中文字幕| 欧美色视频一区免费| 丁香六月欧美| 亚洲精品在线观看二区| 色综合亚洲欧美另类图片| av视频在线观看入口| 欧美中文日本在线观看视频| 岛国在线观看网站| 男人舔女人下体高潮全视频| 中文字幕熟女人妻在线| 男插女下体视频免费在线播放| 色视频www国产| 久久精品亚洲精品国产色婷小说| 亚洲真实伦在线观看| 国产伦人伦偷精品视频| 国产又黄又爽又无遮挡在线| 日韩精品青青久久久久久| 久久亚洲真实| 国产激情偷乱视频一区二区| 精华霜和精华液先用哪个| 成年人黄色毛片网站| 91九色精品人成在线观看| 国产精品av视频在线免费观看| 两性夫妻黄色片| 亚洲专区字幕在线| 美女黄网站色视频| 亚洲欧美激情综合另类| 国产欧美日韩精品一区二区| 中文亚洲av片在线观看爽| 亚洲国产看品久久| 亚洲av电影在线进入| 亚洲第一欧美日韩一区二区三区| 一二三四在线观看免费中文在| 久久中文看片网| 日本一二三区视频观看| 精品午夜福利视频在线观看一区| 亚洲无线观看免费| 美女大奶头视频| 国产成人精品久久二区二区91| 亚洲第一电影网av| 夜夜看夜夜爽夜夜摸| 一二三四社区在线视频社区8| 一级作爱视频免费观看| 99国产精品99久久久久| 亚洲精品在线观看二区| 18禁美女被吸乳视频| 精品99又大又爽又粗少妇毛片 | 丁香欧美五月| 母亲3免费完整高清在线观看| 久久久久久久久免费视频了| 午夜视频精品福利| 叶爱在线成人免费视频播放| 欧美成人免费av一区二区三区| 国产精品一区二区精品视频观看| 一a级毛片在线观看| 精品久久久久久久末码| 日本一本二区三区精品| av在线蜜桃| 国产亚洲欧美在线一区二区| av片东京热男人的天堂| 熟妇人妻久久中文字幕3abv| 国产又色又爽无遮挡免费看| 色精品久久人妻99蜜桃| 久久人妻av系列| 亚洲成人久久性| 久久久精品欧美日韩精品| 欧美成人免费av一区二区三区| 亚洲熟女毛片儿| 国产一区二区在线av高清观看| 免费在线观看成人毛片| 不卡av一区二区三区| 久久精品国产亚洲av香蕉五月| 久久久色成人| 又黄又爽又免费观看的视频| 国产高清视频在线观看网站| 99久久精品国产亚洲精品| 免费在线观看视频国产中文字幕亚洲| 97人妻精品一区二区三区麻豆| 亚洲精品粉嫩美女一区| 99在线视频只有这里精品首页| 中文字幕av在线有码专区| 日韩欧美三级三区| 高清在线国产一区| 草草在线视频免费看| 少妇的丰满在线观看| 日韩欧美 国产精品| 亚洲av日韩精品久久久久久密| 午夜福利免费观看在线| a在线观看视频网站| 国产在线精品亚洲第一网站| 欧美一区二区精品小视频在线| 日韩欧美免费精品| 国产综合懂色| 亚洲 国产 在线| 亚洲精品中文字幕一二三四区| 国产日本99.免费观看| АⅤ资源中文在线天堂| 欧美一级毛片孕妇| 免费一级毛片在线播放高清视频| 麻豆成人av在线观看| 听说在线观看完整版免费高清| www.熟女人妻精品国产| 婷婷精品国产亚洲av| 亚洲欧美日韩无卡精品| 久久精品aⅴ一区二区三区四区| 国产精品美女特级片免费视频播放器 | 99久久国产精品久久久| 熟妇人妻久久中文字幕3abv| 国产三级在线视频| 听说在线观看完整版免费高清| 精品久久久久久久久久久久久| 每晚都被弄得嗷嗷叫到高潮| 欧美性猛交╳xxx乱大交人| 国内久久婷婷六月综合欲色啪| 嫩草影视91久久| 国产成人精品无人区| 免费观看精品视频网站| 男女午夜视频在线观看| 亚洲国产欧美一区二区综合| 天堂网av新在线| 日韩成人在线观看一区二区三区| 免费看十八禁软件| 亚洲 欧美一区二区三区| 18禁黄网站禁片免费观看直播| 欧美乱妇无乱码| 在线十欧美十亚洲十日本专区| 一二三四社区在线视频社区8| 一本久久中文字幕| 成人无遮挡网站| 欧美一级a爱片免费观看看| 男人舔奶头视频| 色播亚洲综合网| av黄色大香蕉| 18禁观看日本| 观看美女的网站| 日本一本二区三区精品| 亚洲成人精品中文字幕电影| 亚洲五月婷婷丁香| 中文资源天堂在线| 岛国视频午夜一区免费看| 国产成人欧美在线观看| 91av网一区二区| 亚洲精品456在线播放app | 国内精品久久久久久久电影| 久久欧美精品欧美久久欧美| 亚洲人成伊人成综合网2020| 村上凉子中文字幕在线| 国产精品亚洲av一区麻豆| 好男人电影高清在线观看| 成人无遮挡网站| 国产成+人综合+亚洲专区| 国产高清视频在线播放一区| 日韩免费av在线播放| 亚洲性夜色夜夜综合| 中出人妻视频一区二区| 日本黄大片高清| 欧美激情在线99| 美女被艹到高潮喷水动态| 一区福利在线观看| 黄色片一级片一级黄色片| 国产精品1区2区在线观看.| 俄罗斯特黄特色一大片| 成人永久免费在线观看视频| 亚洲av第一区精品v没综合| 亚洲最大成人中文| 国产精品久久久久久人妻精品电影| 欧美不卡视频在线免费观看| 麻豆av在线久日| 欧美日韩综合久久久久久 | 免费观看精品视频网站| 欧美日韩中文字幕国产精品一区二区三区| 在线国产一区二区在线| 少妇裸体淫交视频免费看高清| 又黄又粗又硬又大视频| 欧美在线一区亚洲| 亚洲人成网站在线播放欧美日韩| 午夜精品在线福利| 欧美+亚洲+日韩+国产| 亚洲人成伊人成综合网2020| 少妇人妻一区二区三区视频| 午夜福利成人在线免费观看| 香蕉av资源在线| 黑人欧美特级aaaaaa片| 国产三级黄色录像| 香蕉久久夜色| 国产成人精品无人区| 老司机在亚洲福利影院| 桃红色精品国产亚洲av| 老熟妇仑乱视频hdxx| 国产成人av教育| 久久久久亚洲av毛片大全| 夜夜夜夜夜久久久久| 国产成人系列免费观看| 国内久久婷婷六月综合欲色啪| 亚洲片人在线观看| 国产综合懂色| 麻豆国产97在线/欧美| 国产激情偷乱视频一区二区| 国内毛片毛片毛片毛片毛片| 免费无遮挡裸体视频| a级毛片在线看网站| 日韩中文字幕欧美一区二区| 精品一区二区三区视频在线 | 久久久久久国产a免费观看| 人人妻人人看人人澡| 免费大片18禁| 国语自产精品视频在线第100页| 久久天躁狠狠躁夜夜2o2o| 日本成人三级电影网站| 亚洲国产色片| 欧美日本亚洲视频在线播放| 亚洲国产中文字幕在线视频| 岛国视频午夜一区免费看| 国产激情欧美一区二区| 国产蜜桃级精品一区二区三区| 欧美乱码精品一区二区三区| 99热只有精品国产| 亚洲精品乱码久久久v下载方式 | 精品一区二区三区视频在线观看免费| 欧美激情在线99| 在线观看舔阴道视频| 一二三四社区在线视频社区8| 成人18禁在线播放| 日韩成人在线观看一区二区三区| av女优亚洲男人天堂 | 亚洲国产精品成人综合色| 真实男女啪啪啪动态图| 老熟妇仑乱视频hdxx| aaaaa片日本免费| 又紧又爽又黄一区二区| 最近视频中文字幕2019在线8| 午夜福利成人在线免费观看| 一区二区三区高清视频在线| 亚洲av熟女| 欧美日本视频| 亚洲aⅴ乱码一区二区在线播放| 亚洲专区国产一区二区| 亚洲国产色片| 男女午夜视频在线观看| 丰满人妻熟妇乱又伦精品不卡| 国产精品一区二区三区四区久久| 亚洲午夜理论影院| 亚洲黑人精品在线| 人妻夜夜爽99麻豆av| 老司机午夜十八禁免费视频| 精品国产三级普通话版| 中文字幕久久专区| 观看美女的网站| 亚洲欧美精品综合一区二区三区| 国产欧美日韩一区二区精品| 午夜影院日韩av| 蜜桃久久精品国产亚洲av| 亚洲精品色激情综合| 亚洲中文日韩欧美视频| 日韩av在线大香蕉| 免费在线观看影片大全网站| 女人被狂操c到高潮| 日韩欧美一区二区三区在线观看| 免费在线观看日本一区| 搡老妇女老女人老熟妇| 国产亚洲av高清不卡| 国产精品美女特级片免费视频播放器 | 一个人观看的视频www高清免费观看 | 国产精品免费一区二区三区在线| 国产激情欧美一区二区| 国产高清有码在线观看视频| 亚洲精品在线美女| 免费在线观看亚洲国产| 黄色女人牲交| 一区二区三区国产精品乱码| 色播亚洲综合网| 亚洲一区二区三区色噜噜| av片东京热男人的天堂| 免费av毛片视频| 真人做人爱边吃奶动态| 亚洲国产精品成人综合色| 老汉色∧v一级毛片| 日本黄色片子视频| 宅男免费午夜| 国产熟女xx| 成人无遮挡网站| 大型黄色视频在线免费观看| 欧美一级毛片孕妇| 一二三四社区在线视频社区8| 亚洲一区二区三区不卡视频| 高清毛片免费观看视频网站| 日韩欧美国产一区二区入口| 欧美国产日韩亚洲一区| 国产成人福利小说| 国产精品国产高清国产av| 精品欧美国产一区二区三| 亚洲自拍偷在线| 最近最新中文字幕大全电影3| 国产精品爽爽va在线观看网站| 精品福利观看| 国产成人aa在线观看| 国产又色又爽无遮挡免费看| 精品国内亚洲2022精品成人| 91在线精品国自产拍蜜月 | 一二三四社区在线视频社区8| 国产成+人综合+亚洲专区| 十八禁人妻一区二区| 男人舔奶头视频| 国产精品 欧美亚洲| 国产亚洲欧美在线一区二区| 国产乱人伦免费视频| 日韩有码中文字幕| 最新中文字幕久久久久 | 免费电影在线观看免费观看| 亚洲国产日韩欧美精品在线观看 | 久久热在线av| 床上黄色一级片| 人人妻人人看人人澡| 听说在线观看完整版免费高清| 久久中文字幕人妻熟女| 99热这里只有是精品50| 午夜免费观看网址| 熟女少妇亚洲综合色aaa.| 国产一区在线观看成人免费| 欧美3d第一页| 一级a爱片免费观看的视频| 中出人妻视频一区二区| 午夜a级毛片| 亚洲欧洲精品一区二区精品久久久| 久久中文看片网| 黑人欧美特级aaaaaa片| 亚洲精品美女久久av网站| 亚洲欧洲精品一区二区精品久久久| 脱女人内裤的视频| 美女高潮喷水抽搐中文字幕| 亚洲在线观看片| 国产黄a三级三级三级人| 中文字幕av在线有码专区| 露出奶头的视频| 久久精品国产综合久久久| 精品国产乱码久久久久久男人| 国内精品一区二区在线观看| 亚洲五月天丁香| 国产一区二区三区视频了| 国产美女午夜福利| 人人妻人人澡欧美一区二区| 中出人妻视频一区二区| 国产蜜桃级精品一区二区三区| xxx96com| 久久久国产精品麻豆| 757午夜福利合集在线观看| 国产淫片久久久久久久久 | 天堂av国产一区二区熟女人妻| a级毛片在线看网站| 人人妻,人人澡人人爽秒播| 中文字幕最新亚洲高清| 日韩欧美免费精品| 麻豆久久精品国产亚洲av| 校园春色视频在线观看| a级毛片a级免费在线| 亚洲熟女毛片儿| 精品久久蜜臀av无| 亚洲精品国产精品久久久不卡| 免费无遮挡裸体视频| x7x7x7水蜜桃| 国产精品,欧美在线| 亚洲自偷自拍图片 自拍| 别揉我奶头~嗯~啊~动态视频| 久久精品夜夜夜夜夜久久蜜豆| 一区二区三区国产精品乱码| 国产成人影院久久av| 丁香欧美五月| 亚洲av中文字字幕乱码综合| 18禁裸乳无遮挡免费网站照片| 村上凉子中文字幕在线| 国产精品电影一区二区三区| 欧美xxxx黑人xx丫x性爽| 久久久国产欧美日韩av| 窝窝影院91人妻| 在线观看午夜福利视频| 亚洲国产精品sss在线观看| 美女被艹到高潮喷水动态| 成人无遮挡网站| 美女大奶头视频| 真人做人爱边吃奶动态| 99riav亚洲国产免费| 18禁黄网站禁片免费观看直播| 日本三级黄在线观看| 久久精品国产清高在天天线| 桃红色精品国产亚洲av| 日日夜夜操网爽| 成在线人永久免费视频| 欧美精品啪啪一区二区三区| 母亲3免费完整高清在线观看| 在线a可以看的网站| 给我免费播放毛片高清在线观看| 999久久久国产精品视频| 久久这里只有精品19| 久久久久亚洲av毛片大全| 欧美一级毛片孕妇| 香蕉av资源在线| 欧美高清成人免费视频www| 午夜精品在线福利| 国产蜜桃级精品一区二区三区| 老司机深夜福利视频在线观看|