• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Parallel Reinforcement Learning:A Framework and Case Study

    2018-07-31 09:49:56TengLiuMemberIEEEBinTianYunfengAiLiLiFellowIEEEDongpuCaoMemberIEEEandFeiYueWangFellowIEEE
    IEEE/CAA Journal of Automatica Sinica 2018年4期

    Teng Liu,Member,IEEE,Bin Tian,Yunfeng Ai,Li Li,Fellow,IEEE,Dongpu Cao,Member,IEEE,and Fei-Yue Wang,Fellow,IEEE

    Abstract—In this paper,a new machine learning framework is developed for complex system control,called parallel reinforcement learning.To overcome data deficiency of current data-driven algorithms,a parallel system is built to improve complex learning system by self-guidance.Based on the Markov chain(MC)theory,we combine the transfer learning,predictive learning,deep learning and reinforcement learning to tackle the data and action processes and to express the knowledge.Parallel reinforcement learning framework is formulated and several case studies for real-world problems are finally introduced.

    I.INTRODUCTION

    M ACHINE learning especially deep reinforcement learning(DRL)experiences an ultrafast development in recent years[1],[2].No matter in traditional visual detection[3],dexterous manipulation in robotics[4],energy efficiency improvement[5],object localization[6],novel Atari game[7],[8],Leduc poker[9],Doom game[10]and text-based games[11],these data-driven learning approaches show great potential in improving performance and accuracy.However,there are still several issues to impede researchers applying DRL to handle the real complex system problems.

    One of the issues is lack of generalization capability to new goals[3].DRL agents need to collect new data and learn new model parameters for a new target.It is computationally expensive to retrain the learning model.Hence,we need to utilize the limited data well to accommodate the environments via learning.

    Another issue is data inefficiency[8].Acquiring large-scale action and interaction data of real complex systems is arduous.To explore control policy by themselves is very difficult for the learning systems.Thus,it is necessary to create a large number of observations for action and knowledge from the historical available data.

    Finally,the issue is data dependency and distribution.In practical systems,data samples dependency is often uncertain and probability distribution is usually variant.So,it is hard for DRL agents to consider the state,action and knowledge of a learning system in an integrated way.

    In order to address these difficulties,we develop a new parallel reinforcement learning framework for complex system control in this paper.We construct an artificial system analogy to the real system via modelling to constitute a parallel system.Based on the Markov chain(MC)theory,transfer learning,predictive learning,deep learning and reinforcement learning are exhibited to tackle data and action processes and to express knowledge.Furthermore,several application cases of parallel reinforcement learning are introduced to illustrate its usability.It is noticed that the proposed technique in this paper can be regarded as the specification of the parallel learning in[12].

    Fei-Yue Wang first initialized the parallel system theory in 2004[13],[14].In[13]and[14],ACP method was proposed to deal the complex system problem.ACP approach represents artificial societies(A)for modelling,computational experiments(C)for analysis,and parallel execution(P)for control.An artificial system is usually built by modelling,to explore the data and knowledge as the real system does.Through executing independently and complementally in these two systems,the learning model can be more efficient and less data-hungry.ACP approach has been applied in several fields to discuss different problems in complex systems[15]-[17].

    Transfer learning focuses on storing knowledge gained while solving one problem and applying it to a different but related problem.Taking driving cycles of vehicle as an example,we introduce mean traction force(MTF)components to achieve equivalent transformation of them.By transferring limited data via MTF,the generalization capability problem can be relieved.

    Fig.1.Parallel reinforcement learning framework.

    Predictive learning tries to use prior knowledge to build a model of environment by trying out different actions in various circumstances.Taking power demand for example,we introduce fuzzy encoding predictor to forecast the future power demand in different time steps.Based on the MC,historical available data can be used to solve the data inefficiency.

    Deep learning is defined via learning data representations,including multiple layers of nonlinear processing units and the supervised or unsupervised learning of feature representations in each layer.Reinforcement learning is concerned with how agents ought to take actions in an environment so as to maximize some notion of cumulative reward.The main contribution of this paper is combining parallel system with transfer learning,predictive learning,deep learning and reinforcement learning to formulate the parallel reinforcement learning framework to dispose the data dependency and distribution problems in real-world complex systems.

    The rest of this paper is organized as follows.Section II introduces the parallel reinforcement learning framework and relevant components,then several case studies for real-world complex system problems are described in Section III.Finally,we conclude the paper in Section IV.

    II.FRAMEWORK AND COMPONENTS

    A.The Framework and the Parallel System

    The purpose of parallel reinforcement learning is building a closed loop of data and knowledge in the parallel system to determine the next operation in each system,as shown in Fig.1.The data represents the inputs and parameters in artificial and real systems.The knowledge means the records from state space to action space,which we name in the real system as experience and that in the artificial system as policy.The experience can be used to rectify the artificial model and updated policy is utilized to guide the real actor along with feedback from environment.

    Cyber physical systems have attracted increasingly more concerns for their potentials to fuse computational processes with the physical world in the past two decades.Furthermore,cyber-physical-social systems(CPSS)augment the cyber physical system capacity by integrating the human and social characteristics to achieve more effective design and operation[18].The ACP-driven parallel system framework is depicted in Fig.2.The integration of the real and artificial system as a whole is called parallel system.

    In this framework,the physically-defined real system interacts with the software-defined artificial system via three coupled modules within the CPSS.The three modules are control and management,experiment and evaluation,learning and training.The first module belongs to decision maker in these two systems,the second one represents the evaluator and the final one indicates the learning controller.

    ACP=Artificial societies+Computational experiments+Parallel execution.Artificial system is often constructed by descriptive learning based on the observation on the real system due to the development in information and communication technologies.It can help learning controller store more computing results and make more flexible decisions.Thus,the artificial system is parallel to the real system and runs asynchronously to stabilize the learning process and extend the learning capability.

    Fig.2.ACP-driven parallel system framework.

    In the computational experiment stage,the specifications of transfer learning,predictive learning and deep learning are illustrated by the MC theory,as we will discuss them later.For the parallel system,combining these learning processes with reinforcement learning,the parallel reinforcement learning is formulated to derive the experience and policy and to clarify the interaction of them.For a general parallel intelligent system,such knowledge can be applied in different tasks because the learning controller can handle several tasks via rational reasoning[19].

    Finally,parallel execution between the artificial and real systems is expected to enable an optimal operation of these systems[20].Although the artificial system is drawn by the prior data of real system,it will be rectified and improved by the further observation.The consecutive updated knowledge in the artificial system is also used to instruct the real system operation in an efficient way.Owing to the communication of data and knowledge by parallel execution,these two systems can improve by self-guidance.

    B.Transfer Learning

    In this paper,we choose driving cycles as an example to introduce transfer learning,which can be easily popularized for other data in the MC domain.A general driving cycle transformation methodology based on the mean tractive force(MTF)components is introduced in this section.This transformation can convert the existent driving cycles database into an equivalent one with a real MTF value to relieve the data lacking problem.

    MTF is defined as the tractive energy divided by the distance traveled for a whole driving cycle,which is integrated over the entire time interval[0,T]as follows

    where xTis the total distance traveled in a certain driving cycle and calculated as∫v(t)dt,v is the vehicle speed with respect to a certain driving cycle.F is the longitudinal force to propel the vehicle and computed as

    where Fais aerodynamic drag,Fris rolling resistance and Fmis inertial force.ρa(bǔ)is the air density,Cdis the aerodynamic coefficient,and A is the frontal area.Mvis the curb weight,g is the gravitational acceleration,f is the rolling friction coefficient and a is the acceleration.

    The vehicle operating modes are divided as traction,coasting,braking and idling according to the force imposed on the vehicle powertrain[21].Hence,the time interval is depicted as

    where Ttrand Tcoare the traction-mode and coasting-mode regions,respectively.Tbrrepresents the vehicle brakes and Tidis the idling set.

    From(3),it is obvious that the powertrain only provides positive power to wheels in the traction regions.MTF in(1)is specialized as follows:

    Then,MTF components(α,β,γ)are the statistic characteristics measures for a driving cycle that are defined as[22]

    Note that MTF components are related to the speed and acceleration for a specific driving cycle.These measures are employed as the constraints for driving cycle transformation.

    Definition decides MTF is unique for a specific driving cycle,thus inequality and equality constraints are employed to determine the transferred driving cycle.A cost function can be defined by the designer to choose an optimal equivalent one from a set of feasible solutions.This transformation is formulated as a non-linear program(NLP)as

    Fig.3.Transfer learning for driving cycles transformation.

    The purpose of transfer learning is converting historical available data into equivalent one to expand the database.The transferred data is strongly associated with the real environments.Thus,it can be used for generating adaptive control and operations in complex systems,so as to solve the generalization capability and data hungry problems.

    C.Predictive Learning

    Taking power demand of vehicle for example,we introduce predictive learning to forecast the future power demand based on the observed data and processes in parallel system.A better understanding of the real system can then be described and applied to update the artificial system from these new experiences.A power demand prediction technology based on fuzzy encoding predictor(FEP)is illustrated in this section.This approach can also be used to draw more future knowledge from experiences for other parameters in the complex systems.

    Power demand is modelled as a finite-state MC[23]and depicted as Pdem={pj|j=1,...,M}?X,where X?R is bounded.Transition probability of power demand is calculated by maximum likelihood estimator as

    where πijis the transition probability from pito pj.p and p+are the present and next one-step ahead power demands,respectively.Furthermore,Nijindicates the transition count number from pito pj,and Niis the total transition count number initiated from pi.

    All elements πijconstitute the transition probability matrix Π.For fuzzy encoding technique,X is divided into a finite set of fuzzy subsets Φj,j=1,...,M,where Φjis a pair(X,μj(·))and μj(·)is called Lebesgue measurable membership function and defined as

    whereμj(p)reflects the membership degree of p∈X inμj.It is noticed that a continuous state p∈X in the fuzzy encoding may be associated with several states pjof the underlying finite-state MC model[24].

    Two transformations are involved in FEP.The first transformation allocates an M-dimensional possibility(not probability)vector for each p∈X as

    This transformation is named fuzzification and maps power demand in the space X to the vector in M-dimensional possibility vector space?X.Note that it is not necessary for the sum of the elements in possibility vector?O(p)to equal 1.

    The second transformation is called proportional possibilityto-probability transformation,in which the possibility vector ?O(p)is converted into a probability vector O(p)by normalization[23],[24]:

    where this transformation maps?X to an M-dimensional probability vector space, X.The element πijin the transition probability matrix(TPM)Π is interpreted as a transition probability between Φiand Φj.To decode vectors in X back to X,the probability distribution O+(p)is utilized to aggregate the membership functionμ(p)to encode the probability vector of the next state in X:

    The expected value over the possibility vector leads to the next one-step ahead power demand in FEP:

    The centroid and volume of the membership functionμj(p)is expressed as

    Thus,(12)is reformulated as

    where expression(14)is the predicted one-step ahead power demand using FEP.Fig.4 shows an example of predictive learning used for power demand prediction.By doing this,the future power demand of vehicle in different time steps can be determined,and then these data will be used for improving the management and operations in the parallel system by selfguidance.

    Fig.4.Predictive learning for future power demand prediction.

    The goal of predictive learning is generating reasonable data from the prior existed data and real-time observations in the real world.We aim to minimize the differences between real samples and generated samples by tuning the parameters in the predictive learning methodology.Therefore,these generated data are responsible for deriving various experiences and guiding the complex system by learning process,so as to settle the data inefficiency and distribution problem.

    D.Reinforcement Learning

    In the reinforcement learning framework,a learning agent interacts with a stochastic environment.We model the interaction as quintuple(S,A,Π,R,γ),where s∈ S and a∈ A are state variables and control actions sets,Π is the transition probability matrix,r ∈ R is the reward function,and γ∈(0,1)denotes a discount factor.

    The action value function Q(s,a)is defined as the expected reward starting from s and taking the action a:

    The action value function associated with an optimal policy can be found by the Q-learning algorithm as in[25]

    When the state and action space is large,for example the action atconsists of several sub-actions,modelling Q-values Q(s,a)becomes difficult.In this situation,we use both state and action representations as input to a deep neural network to approximate the action value function.

    A deep neural network is composed of an input layer,one or more hidden layers and an output layer.As shown in Fig.5(a),the input vector g=[g1,g2,...,gR]is weighted by elements w1,w2,...,wR,and then summed with a bias b to form the net input n as

    Fig.5.Deep neural network and bidirectional long short-term memory.

    Then,the net input n is affected by an activation function h to generate the neuron output d.

    where activation function usually includes activation function in the hidden layer h1and activation function in the output layer h2.

    In this paper,we propose a bidirectional long short-term memory[26]based deep reinforcement network(BiLSTMDRN)to approximate the action value function in reinforcement learning,see Fig.5(b)for an illustration.This structure consists of a pair of deep neural networks,one for state variable stembedding and the other for control sub-actionsembeddings.As the bidirectional LSTM has a larger capacity due to its nonlinear structure,we expect it will capture more details on how the embeddings in each sub-action are combined into an action embedding.Finally,a pairwise interaction function(e.g.,inner product)is used to compute new Q(st,at)via combining the state and sub-actions neuron output as

    where K is the number of the sub-actions,and Q(st,)represents the expected accumulated future rewards by including this sub-action.

    Combining the ideas of parallel system,transfer learning,predictive learning and reinforcement learning,we can formulate a closed loop of data and knowledge,named parallel reinforcement learning,as described in Fig.1.Several case studies for real-world complex system problems are introduced and discussed in the next section.

    III.CASE STUDIES OF PARALLEL REINFORCEMENT LEARNING

    A.Existing Case Studies in Parallel Reinforcement Learning Framework

    Parallel reinforcement learning serves as a reasonable and suitable framework to analyse the real world complex system.It consists of a self-boosting process in the parallel system,a self-adaptive process by transfer learning,a self-guided process by predictive learning and big data screening and generating process by BiLSTM-DRN.Learning process becomes more efficient and continuous in the parallel reinforcement learning framework.

    Several complex systems have been researched and analysed in the perspective of parallel reinforcement learning,such as transportation systems[27],[28],and vision systems[29].A traffic flow prediction system was designed in[27],which considered the spatial and temporal correlations inherently.First,an artificial system named stacked autoencoder model was built to learn generic traffic flow features.Second,the synthetic data were trained by a layer-wise greedy method in the deep learning architecture.Finally,predictive learning was used to achieve traffic flow prediction and self-guidance for the parallel system.A survey on the development of the data-driven intelligent transportation system(D-DITS)was introduced in[28].The functionality of D-DITS’s key components and some deployment issues associated with its future research were addressed in[28].

    Also,a parallel reinforcement learning framework has also been applied to address the problems in visual perception and understanding[29].To draw an artificial vision system based on the observations from real scenes,the synthetic data can be used for feature analysis,object analysis and scene analysis.This novel research methodology,named parallel vision,was proposed for perception and understanding of complex scenes.

    Furthermore,autonomous learning system for vehicle energy efficiency improvement in[30]can also be put into parallel reinforcement learning framework.First,a plug-in hybrid electric vehicle was imitated to construct the parallel system.Then,historical driving record for the real vehicle was collected to learn autonomously the optimal fuel use via a deep neural network and reinforcement learning.Finally,this trained policy can guide the real vehicle operations and improve control performance.A better understanding of the real vehicle can then be obtained and used to adjust the artificial system from these new experiences.

    B.New Applications Using Parallel Reinforcement Learning Methods

    Recently,we designed a driving cycles transformation based adaptive energy management system for a hybrid electric vehicle(HEV).There exist two major difficulties in the energy management problem of HEV.First,most of energy management strategies or predefined rules cannot adapt to changing driving conditions.Second,model-based approaches used in energy management require accurate vehicle models,which bring a considerable model parameter calibration cost.Hence,we apply parallel reinforcement learning framework into the energy management problem of HEV,as depicted in Fig.6.More precisely,the core idea of this methodology is bi-level.

    The up-level characterizes how to transform driving cycles using transfer learning by considering the induced matrix norm(IMN).Specially,TPM of power demand are computed and IMN is employed as a critical criterion to identify the differences of TPMs and to determine the alteration of control strategy.The lower-level determines how to set the corresponding control strategies with the transferred driving cycle by using model-free reinforcement learning algorithm.In other words,we simulate the HEV as an artificial system to sample the possible energy management solutions,use transfer learning to make the computed strategies adaptive to real world driving conditions,and use reinforcement learning to generate the corresponding controls.Tests demonstrate that the proposed strategy exceeds the conventional reinforcement learning approach in both calculation speed and control performance.

    Fig.6.Parallel reinforcement learning for energy management of HEV.

    Fig.7.Parallel reinforcement learning for energy efficiency of HTV.

    Furthermore,we construct an energy efficiency improvement system in parallel reinforcement learning framework for a hybrid tracked vehicle(HTV).Specifically,we combine the simulated artificial vehicle with real vehicle to constitute the parallel system,use predictive learning to realize power demand prediction for further self-guidance and use reinforcement learning for control policy calculation.This approach also includes two layers,see Fig.7 for a visualization of such idea.The first layer discusses how to accurately forecast the future power demand using FEP based on the MC theory.Kullback-Leibler(KL)divergence rate is employed to decide the differences of TPMs and updating of control strategy.The second layer computes the relevant control policy based on the predicted power demand and reinforcement learning technique.Finally,comparison shows that the proposed control policy is superior to the primary reinforcement learning approach in energy efficiency improvement and computational speed.

    In the future,we plan to apply BiLSTM-DRN to process and train the large real vehicle data for optimal energy management strategy computation.The objective is to realize realtime control using the parallel reinforcement learning method in our self-made tracked vehicle.More importantly,we will apply parallel reinforcement learning framework into multimissions of automated vehicles[30],such as decision making,trajectory planning and so on.To address the existing disadvantages of traditional data-driven methods,we expect that parallel reinforcement learning can promote the development of machine learning.

    IV.CONCLUSION

    The general framework and case studies of parallel reinforcement learning for complex systems are introduced in this paper.The purpose is to build a closed loop of data and knowledge in the parallel system to guide the real system operation or improve the artificial system precision.Particularly,ACP approach is used to construct the parallel system that contains an artificial system and a real system.Transfer learning is utilized to achieve driving cycle transformation by mean of tractive force components.Predictive learning is applied to forecast the future power demand via fuzzy encoding predictor.To train data in the large action and state space,we introduce BiLSTM-DRN to approximate the action value function in reinforcement learning.

    Data-driven models are usually viewed as a component irrelevant to the data in learning process,which results in the largescale exploration and observation-insufficiency problems.Furthermore,data in these models tend to be inadequate,and the general principle to organize these models remains absent.By combining parallel system,transfer learning,predictive learning,deep learning and reinforcement learning,we believe that parallel reinforcement learning can effectively address these problems and promote the development of machine learning.

    久久久a久久爽久久v久久| 最新中文字幕久久久久| 日本猛色少妇xxxxx猛交久久| 狂野欧美激情性xxxx在线观看| 有码 亚洲区| 中国国产av一级| 色视频在线一区二区三区| 久久国内精品自在自线图片| 又黄又爽又刺激的免费视频.| 欧美激情国产日韩精品一区| 男女边摸边吃奶| 精品午夜福利在线看| 最后的刺客免费高清国语| 欧美成人精品欧美一级黄| 日日摸夜夜添夜夜添av毛片| 亚洲av欧美aⅴ国产| 日韩大片免费观看网站| 午夜福利视频1000在线观看| 天堂网av新在线| 国产老妇女一区| 国内揄拍国产精品人妻在线| 欧美成人一区二区免费高清观看| 最近最新中文字幕大全电影3| 成年版毛片免费区| 国产成人91sexporn| 欧美zozozo另类| 亚洲精品456在线播放app| 免费看不卡的av| 午夜福利高清视频| 久久ye,这里只有精品| 人妻少妇偷人精品九色| 亚洲精华国产精华液的使用体验| 久久99蜜桃精品久久| 国产亚洲最大av| 男人爽女人下面视频在线观看| 观看免费一级毛片| 人体艺术视频欧美日本| 国产成年人精品一区二区| 国产在视频线精品| 一级爰片在线观看| 国产女主播在线喷水免费视频网站| 免费电影在线观看免费观看| 欧美成人一区二区免费高清观看| 亚洲在线观看片| 精品国产一区二区三区久久久樱花 | 内射极品少妇av片p| 亚洲真实伦在线观看| 日韩强制内射视频| 亚洲精品中文字幕在线视频 | 午夜激情福利司机影院| 美女被艹到高潮喷水动态| 永久免费av网站大全| 777米奇影视久久| 亚洲最大成人手机在线| 亚洲在线观看片| 永久网站在线| 亚洲精品国产成人久久av| 精品人妻熟女av久视频| 人人妻人人爽人人添夜夜欢视频 | 日日摸夜夜添夜夜添av毛片| 精华霜和精华液先用哪个| 黄色怎么调成土黄色| 婷婷色av中文字幕| 网址你懂的国产日韩在线| 少妇丰满av| 国产成人免费无遮挡视频| 久久久久久久大尺度免费视频| 久久久久久久久久成人| 菩萨蛮人人尽说江南好唐韦庄| 国产免费视频播放在线视频| 国产精品无大码| 国产精品精品国产色婷婷| 国产乱来视频区| 亚洲欧美成人精品一区二区| 99热国产这里只有精品6| 亚洲欧美日韩卡通动漫| 丝袜喷水一区| 边亲边吃奶的免费视频| 一级毛片久久久久久久久女| 如何舔出高潮| 精品熟女少妇av免费看| 亚洲内射少妇av| 国产人妻一区二区三区在| 可以在线观看毛片的网站| 国产乱人视频| 超碰av人人做人人爽久久| 免费看av在线观看网站| 久久久久国产网址| 最近最新中文字幕免费大全7| 婷婷色综合www| 免费少妇av软件| 在线看a的网站| 国产男女内射视频| 免费看日本二区| 免费看光身美女| 日韩中字成人| 日韩欧美精品免费久久| 精品久久久久久久人妻蜜臀av| 日韩一本色道免费dvd| 精品一区在线观看国产| 干丝袜人妻中文字幕| 日韩,欧美,国产一区二区三区| 3wmmmm亚洲av在线观看| 毛片女人毛片| 国产精品一区二区三区四区免费观看| 国产精品国产三级国产av玫瑰| 日韩亚洲欧美综合| 久热这里只有精品99| 日产精品乱码卡一卡2卡三| 亚洲最大成人av| 国产69精品久久久久777片| 看黄色毛片网站| 国产极品天堂在线| 色播亚洲综合网| av线在线观看网站| 国内少妇人妻偷人精品xxx网站| 久久精品综合一区二区三区| 性色av一级| 人妻制服诱惑在线中文字幕| 又粗又硬又长又爽又黄的视频| 午夜爱爱视频在线播放| 国产精品99久久久久久久久| 看非洲黑人一级黄片| videossex国产| 在线观看一区二区三区激情| 自拍偷自拍亚洲精品老妇| 国产黄片视频在线免费观看| 免费黄网站久久成人精品| .国产精品久久| 国产一区亚洲一区在线观看| 亚洲成人久久爱视频| 免费av观看视频| 精品人妻视频免费看| 亚洲欧美日韩卡通动漫| 欧美97在线视频| 亚洲一区二区三区欧美精品 | 欧美区成人在线视频| 国产精品人妻久久久影院| 亚洲成人久久爱视频| 欧美+日韩+精品| 九九久久精品国产亚洲av麻豆| 18禁裸乳无遮挡免费网站照片| 大又大粗又爽又黄少妇毛片口| 亚洲精品乱码久久久久久按摩| 一级二级三级毛片免费看| 亚洲精品456在线播放app| 欧美日韩综合久久久久久| 大片电影免费在线观看免费| 人妻系列 视频| 天天一区二区日本电影三级| 国产成人91sexporn| 久久久久精品性色| 黄色一级大片看看| 国产国拍精品亚洲av在线观看| 在线免费观看不下载黄p国产| 欧美+日韩+精品| 菩萨蛮人人尽说江南好唐韦庄| 成年人午夜在线观看视频| 看黄色毛片网站| 欧美bdsm另类| 精品人妻熟女av久视频| 夫妻性生交免费视频一级片| 嫩草影院精品99| 国产爽快片一区二区三区| 亚洲欧美精品自产自拍| 久久6这里有精品| 观看免费一级毛片| av免费观看日本| 欧美xxxx黑人xx丫x性爽| 国产淫语在线视频| 26uuu在线亚洲综合色| 亚洲欧美清纯卡通| 精品少妇黑人巨大在线播放| 精品午夜福利在线看| 天堂俺去俺来也www色官网| 国产毛片a区久久久久| 国产精品久久久久久精品电影| 丝袜脚勾引网站| 内射极品少妇av片p| 日日撸夜夜添| 少妇 在线观看| 国产精品.久久久| 中文精品一卡2卡3卡4更新| 狂野欧美白嫩少妇大欣赏| 人妻制服诱惑在线中文字幕| 一级毛片电影观看| 涩涩av久久男人的天堂| 国产 一区精品| 婷婷色综合www| 在线播放无遮挡| 亚洲美女视频黄频| 建设人人有责人人尽责人人享有的 | av专区在线播放| av国产免费在线观看| 国内精品宾馆在线| 国产毛片a区久久久久| 又粗又硬又长又爽又黄的视频| 国产69精品久久久久777片| 欧美日韩综合久久久久久| 综合色av麻豆| 真实男女啪啪啪动态图| 免费观看a级毛片全部| 精品久久久精品久久久| 一个人看视频在线观看www免费| 爱豆传媒免费全集在线观看| 欧美丝袜亚洲另类| 国产免费福利视频在线观看| 超碰av人人做人人爽久久| 中文在线观看免费www的网站| 午夜日本视频在线| 亚洲精品影视一区二区三区av| 国产淫语在线视频| 七月丁香在线播放| 色视频www国产| 国产av码专区亚洲av| 精品一区在线观看国产| 亚洲av成人精品一区久久| 日本与韩国留学比较| 精品久久久精品久久久| 岛国毛片在线播放| 一级二级三级毛片免费看| 在线观看人妻少妇| 视频区图区小说| 狂野欧美激情性bbbbbb| 永久网站在线| 精品久久久久久久末码| 国产精品99久久久久久久久| 一级毛片电影观看| 国产免费又黄又爽又色| www.色视频.com| 午夜激情福利司机影院| 国产有黄有色有爽视频| 色视频在线一区二区三区| 黄色欧美视频在线观看| 少妇人妻久久综合中文| 少妇熟女欧美另类| 亚洲av日韩在线播放| 精品少妇久久久久久888优播| 天堂网av新在线| 国产一区二区在线观看日韩| 性色av一级| 亚洲欧美一区二区三区黑人 | 国产av不卡久久| 久久久久久久大尺度免费视频| 交换朋友夫妻互换小说| 能在线免费看毛片的网站| 国产精品国产av在线观看| 少妇人妻久久综合中文| 亚洲精品国产av蜜桃| av网站免费在线观看视频| 午夜福利视频精品| 亚洲国产日韩一区二区| 搡老乐熟女国产| 真实男女啪啪啪动态图| 欧美日韩精品成人综合77777| 久久久久精品性色| 99视频精品全部免费 在线| 禁无遮挡网站| 免费大片黄手机在线观看| 精品久久国产蜜桃| 日韩视频在线欧美| 亚洲怡红院男人天堂| 99热6这里只有精品| 人体艺术视频欧美日本| 青青草视频在线视频观看| 精品久久久噜噜| 精品国产三级普通话版| 香蕉精品网在线| 国产永久视频网站| 王馨瑶露胸无遮挡在线观看| 日韩国内少妇激情av| 2018国产大陆天天弄谢| 在线观看一区二区三区激情| av免费在线看不卡| 99久久人妻综合| 久久久亚洲精品成人影院| 一区二区三区免费毛片| 久久鲁丝午夜福利片| 日本猛色少妇xxxxx猛交久久| 极品少妇高潮喷水抽搐| 精品久久久久久电影网| 欧美高清性xxxxhd video| 久久99蜜桃精品久久| av女优亚洲男人天堂| 久久久久久久久久人人人人人人| 一区二区三区四区激情视频| 久久久欧美国产精品| 97超视频在线观看视频| 极品少妇高潮喷水抽搐| 国产精品久久久久久精品电影| 嘟嘟电影网在线观看| 日韩av在线免费看完整版不卡| 国语对白做爰xxxⅹ性视频网站| 欧美激情国产日韩精品一区| 99九九线精品视频在线观看视频| av专区在线播放| 日韩大片免费观看网站| 乱码一卡2卡4卡精品| 少妇被粗大猛烈的视频| 国产一级毛片在线| av在线观看视频网站免费| 麻豆成人午夜福利视频| 纵有疾风起免费观看全集完整版| 热99国产精品久久久久久7| .国产精品久久| 91在线精品国自产拍蜜月| 18+在线观看网站| 2018国产大陆天天弄谢| 联通29元200g的流量卡| 午夜激情久久久久久久| 在线观看一区二区三区激情| 国产欧美另类精品又又久久亚洲欧美| tube8黄色片| 国内揄拍国产精品人妻在线| 婷婷色av中文字幕| 亚洲欧美日韩无卡精品| 欧美3d第一页| 91aial.com中文字幕在线观看| 精品人妻熟女av久视频| 天天躁夜夜躁狠狠久久av| 日本一二三区视频观看| 最近中文字幕2019免费版| 啦啦啦中文免费视频观看日本| 午夜免费男女啪啪视频观看| 欧美精品国产亚洲| 在线观看一区二区三区激情| 大码成人一级视频| 亚洲人成网站高清观看| 一本色道久久久久久精品综合| 九九爱精品视频在线观看| 91aial.com中文字幕在线观看| 亚洲精品成人av观看孕妇| 1000部很黄的大片| 亚洲国产色片| 久久久久久九九精品二区国产| 毛片女人毛片| av在线蜜桃| 禁无遮挡网站| 大话2 男鬼变身卡| 在线免费十八禁| 寂寞人妻少妇视频99o| 性色avwww在线观看| 久久久久久久久久久免费av| 亚洲婷婷狠狠爱综合网| 大陆偷拍与自拍| 国产成人免费无遮挡视频| 日本熟妇午夜| 国产av国产精品国产| 久久久久久久久久成人| 久久精品久久精品一区二区三区| 九草在线视频观看| 91在线精品国自产拍蜜月| 久久久久久久久久成人| 久久精品久久精品一区二区三区| 色综合色国产| 美女内射精品一级片tv| 99视频精品全部免费 在线| 国模一区二区三区四区视频| 熟女电影av网| 一本一本综合久久| 久久久久久九九精品二区国产| 在线 av 中文字幕| 天天躁夜夜躁狠狠久久av| 丝袜脚勾引网站| 欧美精品一区二区大全| 国产人妻一区二区三区在| 交换朋友夫妻互换小说| 精品人妻熟女av久视频| 日日啪夜夜撸| 啦啦啦啦在线视频资源| 日本与韩国留学比较| 日本黄色片子视频| 亚洲精品日韩av片在线观看| 久热久热在线精品观看| 成人黄色视频免费在线看| 色视频www国产| 黄片无遮挡物在线观看| 亚洲精品456在线播放app| 哪个播放器可以免费观看大片| 日本黄色片子视频| 精品久久久久久久久亚洲| 中文字幕亚洲精品专区| 亚洲欧洲国产日韩| 新久久久久国产一级毛片| 久久久久久伊人网av| av黄色大香蕉| 国产成年人精品一区二区| 午夜精品一区二区三区免费看| 中文精品一卡2卡3卡4更新| 夜夜爽夜夜爽视频| 久久久精品94久久精品| 亚洲性久久影院| 久久人人爽av亚洲精品天堂 | 97人妻精品一区二区三区麻豆| 国产精品久久久久久久久免| 亚洲欧洲国产日韩| 乱码一卡2卡4卡精品| 国产精品国产三级国产av玫瑰| 久久久久久久亚洲中文字幕| 夫妻午夜视频| 麻豆成人av视频| 观看美女的网站| 欧美人与善性xxx| 国产高清有码在线观看视频| 白带黄色成豆腐渣| 午夜福利视频1000在线观看| av网站免费在线观看视频| 最近的中文字幕免费完整| 汤姆久久久久久久影院中文字幕| 国产一级毛片在线| 草草在线视频免费看| av线在线观看网站| xxx大片免费视频| 男人添女人高潮全过程视频| 噜噜噜噜噜久久久久久91| 麻豆成人av视频| 国产高清有码在线观看视频| 三级男女做爰猛烈吃奶摸视频| 男插女下体视频免费在线播放| 身体一侧抽搐| 97精品久久久久久久久久精品| 日日啪夜夜撸| 久久久久久久大尺度免费视频| 视频区图区小说| 中文字幕人妻熟人妻熟丝袜美| 国产成年人精品一区二区| 成年女人在线观看亚洲视频 | 国产精品一区二区三区四区免费观看| 日韩欧美精品免费久久| 中文天堂在线官网| 日本黄色片子视频| 亚洲欧美成人综合另类久久久| 伊人久久国产一区二区| 国产成年人精品一区二区| 秋霞伦理黄片| 欧美日韩视频高清一区二区三区二| 久久精品国产亚洲av涩爱| 久久精品久久精品一区二区三区| 99精国产麻豆久久婷婷| 国产精品一区www在线观看| 国产极品天堂在线| 国产欧美日韩一区二区三区在线 | 免费少妇av软件| 久久久精品94久久精品| 亚洲国产日韩一区二区| 一本久久精品| 国产乱人视频| 欧美成人精品欧美一级黄| 亚洲伊人久久精品综合| 欧美一级a爱片免费观看看| 九色成人免费人妻av| 日韩成人伦理影院| 久久韩国三级中文字幕| 国内精品美女久久久久久| 丝瓜视频免费看黄片| 中国三级夫妇交换| 亚洲人成网站在线观看播放| 天美传媒精品一区二区| 人体艺术视频欧美日本| 日韩 亚洲 欧美在线| av在线观看视频网站免费| 国产精品99久久99久久久不卡 | 免费在线观看成人毛片| 最新中文字幕久久久久| 午夜精品一区二区三区免费看| 精品一区在线观看国产| 大香蕉久久网| 日本一二三区视频观看| 久久久午夜欧美精品| 亚洲国产成人一精品久久久| 日韩三级伦理在线观看| 免费看不卡的av| 欧美激情在线99| 精品国产露脸久久av麻豆| 国产老妇伦熟女老妇高清| 午夜日本视频在线| 亚洲久久久久久中文字幕| 黑人高潮一二区| 人妻 亚洲 视频| 超碰av人人做人人爽久久| 免费av不卡在线播放| 国产片特级美女逼逼视频| 2021天堂中文幕一二区在线观| 天堂俺去俺来也www色官网| 在线观看av片永久免费下载| 亚洲欧美中文字幕日韩二区| 亚洲精品aⅴ在线观看| 国产黄色视频一区二区在线观看| 国产一区二区三区av在线| 97在线人人人人妻| 亚洲国产欧美人成| www.av在线官网国产| 国产亚洲av嫩草精品影院| 中文字幕av成人在线电影| 免费看av在线观看网站| 丰满人妻一区二区三区视频av| 人妻 亚洲 视频| 一区二区三区四区激情视频| 日本黄大片高清| 国产精品一二三区在线看| 久久99热这里只有精品18| 男人添女人高潮全过程视频| 天美传媒精品一区二区| 免费观看在线日韩| 性色av一级| 国产成人免费观看mmmm| 欧美高清成人免费视频www| 国产成人a∨麻豆精品| 国产免费福利视频在线观看| 干丝袜人妻中文字幕| 成人国产av品久久久| 日韩大片免费观看网站| 亚洲精品一二三| 一区二区三区免费毛片| 午夜日本视频在线| 亚洲图色成人| 欧美另类一区| 人人妻人人澡人人爽人人夜夜| 简卡轻食公司| 最新中文字幕久久久久| 国产欧美日韩精品一区二区| 国产成人免费无遮挡视频| 午夜福利网站1000一区二区三区| 欧美成人精品欧美一级黄| 五月玫瑰六月丁香| 精品少妇黑人巨大在线播放| 久久国内精品自在自线图片| 欧美日韩在线观看h| 真实男女啪啪啪动态图| 在线观看美女被高潮喷水网站| 三级男女做爰猛烈吃奶摸视频| 51国产日韩欧美| 搡女人真爽免费视频火全软件| 一级爰片在线观看| 日本av手机在线免费观看| 永久网站在线| 晚上一个人看的免费电影| 国产成人精品婷婷| 久久久久国产网址| 亚洲国产色片| 视频区图区小说| 精品久久久久久久人妻蜜臀av| 天天躁日日操中文字幕| 亚洲av欧美aⅴ国产| 国产爽快片一区二区三区| 男女边吃奶边做爰视频| 网址你懂的国产日韩在线| 狂野欧美白嫩少妇大欣赏| 草草在线视频免费看| 国产精品国产av在线观看| videossex国产| 国产精品国产三级专区第一集| 国产久久久一区二区三区| 久久精品综合一区二区三区| 欧美最新免费一区二区三区| 久久久久性生活片| 99热网站在线观看| 在线播放无遮挡| 春色校园在线视频观看| 久久国产乱子免费精品| 青春草视频在线免费观看| 日日摸夜夜添夜夜爱| 午夜福利高清视频| 直男gayav资源| 日韩强制内射视频| 全区人妻精品视频| 啦啦啦啦在线视频资源| 自拍欧美九色日韩亚洲蝌蚪91 | 日韩一区二区视频免费看| 男男h啪啪无遮挡| 亚洲av国产av综合av卡| 国产日韩欧美亚洲二区| 亚洲av一区综合| 成人毛片60女人毛片免费| 亚洲久久久久久中文字幕| av一本久久久久| 在线天堂最新版资源| 免费看日本二区| 日本猛色少妇xxxxx猛交久久| 一个人看的www免费观看视频| 国产中年淑女户外野战色| 国产精品久久久久久精品电影| 91狼人影院| 成年人午夜在线观看视频| 少妇人妻 视频| 欧美高清成人免费视频www| 美女cb高潮喷水在线观看| av国产免费在线观看| 久久久精品免费免费高清| 大片免费播放器 马上看| 国产精品偷伦视频观看了| 成人综合一区亚洲| 亚洲国产精品成人综合色| 婷婷色麻豆天堂久久| 国产高清三级在线| 黄色怎么调成土黄色| 欧美精品人与动牲交sv欧美| 丝瓜视频免费看黄片| 18禁在线播放成人免费| 两个人的视频大全免费| 亚洲av在线观看美女高潮| 一级毛片久久久久久久久女| 永久免费av网站大全| 久久久久久久久久久丰满| 在线观看美女被高潮喷水网站| 国产高潮美女av| av黄色大香蕉| 69人妻影院| 男女啪啪激烈高潮av片| 精品人妻一区二区三区麻豆| 欧美日韩视频高清一区二区三区二| 亚洲精品aⅴ在线观看| 又爽又黄a免费视频| 国产黄片视频在线免费观看| 久久精品人妻少妇|