• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Analytical redundancy of variable cycle engine based on variable-weights neural network

    2022-11-13 07:30:58ZihaoZHANGXianghuaHUANGTianhongZHANG
    CHINESE JOURNAL OF AERONAUTICS 2022年10期

    Zihao ZHANG, Xianghua HUANG, Tianhong ZHANG

    Jiangsu Province Key Laboratory of Aerospace Power System, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China

    KEYWORDS Analytical redundancy;Degradation;Multiple variables;Neural networks;Variable cycle engine

    Abstract In this paper, variable-weights neural network is proposed to construct variable cycle engine’s analytical redundancy, when all control variables and environmental variables are changing simultaneously, also accompanied with the whole engine’s degradation. In another word,variable-weights neural network is proposed to solve a multi-variable, strongly nonlinear,dynamic and time-varying problem. By making weights a function of input, variable-weights neural network’s nonlinear expressive capability is increased dramatically at the same time of decreasing the number of parameters. Results demonstrate that although variable-weights neural network and other algorithms excel in different analytical redundancy tasks, due to the fact that variableweights neural network’s calculation time is less than one fifth of other algorithms, the calculation efficiency of variable-weights neural network is five times more than other algorithms. Variableweights neural network not only provides critical variable-weights thought that could be applied in almost all machine learning methods,but also blazes a new way to apply deep learning methods to aeroengines.

    1. Introduction

    Analytical redundancy can be used to provide reference signal for diagnostics or reconstruction signal in fault tolerance control. Analytical redundancy is usually defined as signal calculated by algorithms or methods rather than signals measured by physical sensors. For example, on the premise of having installed mass flow sensors and pressure sensors at entrance and exit of compressors,when aeroengine’s rotation speed sensor gets failed, then by interpolating compressor’s characteristics map using mass flow and pressure ratio, rotation speed could be identified uniquely. Rotation speed signal calculated by interpolation is rotation speed analytical redundancy. For diagnostics, analytical redundancy could be used as reference signal to vote whether faults occur.1In fault tolerance control,analytical redundancy could be used as alternative feedback signal of fault sensors so that the close loop of control system can keep intact and finish control task.2

    The interpolation method mentioned in the first paragraph is the most classical, primitive and original method to get analytical redundancy. The advantage of this method is reliable and simple to use, but its disadvantage is also manifested─limited to characteristics maps. Aeroengine’s components characteristics maps are plotted when aeroengines operate in steady status. For example, when rotator speed of aeroengines dramatically increases or decreases, aeroengines’dynamic working points do not exist in steady characteristics maps. Therefore, dynamic analytical redundancy cannot be provided in dynamic process.

    To break through the limitation of steady status, scientists and engineers replace aeroengines’ math model’s Newton-Raphson iteration with dynamics, like volume dynamics,3rotor dynamics4and multibody dynamics,5by which means not only the calculation speed is drastically increased but also the simulated dynamic process gets more approximate to the real physical process. Theoretically, with the help of dynamic models, as long as entrance conditions and control variables are ascertained,all status parameters like rotation speed,pressure ratio, and temperature could be acquired, no matter in steady status or dynamic process. However, in practice, even if the principle of similitude has functioned a lot, when aeroengines operate in different working points,6a new set of modification coefficients have to be applied to aeroengines so that model’s nonlinear characteristics could keep consistent with actual aeroengines.The process of acquiring and applying sets of modification coefficients to models is usually known as model modification.7Model modification does address the problem of constructing dynamic analytical redundancy within a wide working range, but a new problem brought by these coefficients is numerous modification coefficients to be ascertained manually. Other model-based derivative methods, like Kalman filter and extended Kalman filter,8also suffer from these to-be-ascertained modification coefficients.

    Manually adjusting a vast amount of parameters is a great burden to engineers, so a question is raised─could all parameters get adjusted automatically? This question is exactly the core thought of machine leaning─automatically adjusting parameters.

    Numerous machine leaning methods have been proposed to construct analytical redundancy, representatives of which are Support Vector Machine (SVM) and Extreme Learning Machine (ELM). Zhao and Sun proposed an SVM to construct analytical redundancy.9By adopting greedy stagewise and iterative strategies, the SVM is capable of online estimating parameters of complicated systems. Zhou et al. proposed an ELM to construct analytical redundancy for sensor fault diagnostics.10By selectively updating output weights of neural networks according to prediction accuracy and norms of output weight vectors, the prediction capability of the ELM is enhanced.

    However,it can be seen that whether in Zhao’s experiments or Zhou’s simulations, the harshest situation is dynamic process only involving three control variables, with two components’ degradation, and always at 0 height and 0 Mach number. A question is naturally proposed─why don’t these methods consider more control variables,more points in flight envelope, and the whole engine’s degradation? It is because SVM and ELM belong to shallow learning methods, a category of methods that do not have enough nonlinear expressive capability, which also explains why most shallow learning methods have to be online or onboard, because their parameters have to be updated to adapt to aeroengine’s nonlinear characteristics at different working points. From the perspective of changing parameters, shallow learning methods are not different with model modification, but it can be observed that the key point to construct strongly nonlinear systems’analytical redundancy is increasing algorithms’ nonlinear expressive capability.

    Deep learning algorithms are exactly the production of solving strongly nonlinear problems, and their representatives are Long Short Time Memory neural network (LSTM) and Convolutional Neural Network (CNN). LSTM is a method that takes advantage of forget-and-memory gates to decide that sequence data at which moment should be assigned greater weights, which is the main source of its nonlinear expressive capability.11Nevertheless, the mechanism of forget-and-memory gates narrows the application field of LSTM to sequence problems,like natural language processing,

    where input is a row vector. CNN originates from image processing problem, but is widely used in language processing,object detection, classification and regression. This is because CNN could incorporate almost all nonlinear layers, like activation, normalization, pooling, or even LSTM, which means that its topological structure could be extremely complicated.Exactly relying on complex topological structures, CNN can express complex system and solve strongly nonlinear problems.12Yet, the subsequent problem is the huge amounts of parameters and slow calculation speed, which is unacceptable to objects with requirement of real-time calculation, like aeroengines.

    Naturally, a question is proposed: how to increase algorithms’ nonlinear expressive capability at the same time of decreasing parameters?For now,the answer could be pruning methods,13graphic CNNs,14or parameters’ dimensional reduction,15but thinking more deeply, what is the key factor to determine an algorithm’s nonlinear expressive capability?The answer to this question could be the key to unlock the contradiction between huge amounts of parameters and nonlinear expressive capability.

    This paper offers a new perspective to explore the nature of nonlinear expressive capability. Inspired by a simple phenomenon, that is for a simple function f(x)=x2, if rewriting the function into f(x)=w(x)·x+b, which is the basic form of neural networks, it can be found that, when b=0,w(x)=x, the function could be perfectly fitted. However, by checking all neural networks, after training, weights are constants instead of a function of input.Thus,we generalize an idea that making weights a function of input might be the key factor to improve algorithms’nonlinear expressive capability.

    This paper is organized as follows.In Section 2,Section 2.1 gives the definition of symbols and operations used in this paper, and Section 2.2 introduces the structure of Variable-Weights Neural Networks (VWNNs) and the core algorithm- variable-weights algorithm. In Section 3, experiments are conducted to demonstrate the performance of VWNN. In order to make analytical redundancy based on VWNN more convincing and compatible with milder situation, the experiment object is chosen as Variable Cycle Engine (VCE), and the experiment situation is that all environmental variables and control variables change simultaneously, also accompanied as the whole engine’s degradation. Section 3.1 is a brief introduction to VCE and data set.Steady and dynamic performance of VWNN is demonstrated in Section 3.2 and 3.3 respectively. Section 3.4 compares calculation efficiency between VWNN and other methods. Section 4 concludes the paper briefly.

    2. Method

    2.1. Definition of symbols and operations

    Throughout this paper,research objects involve arrays of different dimensionality. Consequently, we refer to one-dimension arrays as vectors, two-dimension arrays as matrices, and three-dimension arrays as tensors.Bold lower-case letters(e.g.a)denote vectors,column vector by default.airepresents the element of a at i row.Bold upper-case letters(e.g.A)are matrices,and calligraphic bold upper-case letters(e.g.A)are tensors.

    The first dimension of all arrays is row, the second dimension is column, and the third dimension is page. Besides, for convenience of transforming the algorithm into codes, getting replicated and accelerating calculation,all formulas are finally transformed into operation among vectors, matrices and tensors by using operation defined below.

    Operators ‘‘.+” ‘‘.-” ‘‘.×” ‘‘.^” denote operation by element. If dimensions of two arrays do not match, the lowdimension array will be expanded into higher dimension by being replicated following the lost dimension, as shown in

    2.2. Algorithm

    As shown in Fig. 1, VWNN includes a core algorithm—vari able-weights layer,and accessory algorithms—input layer,output layer and feature separation layer. Before formulas, it is noted that, during forward propagation, input of every layer is output of former layer. For back propagation, output loss of each layer is input loss of latter layer. For example, output loss of activation layer is input loss of variable-weights layer.

    (1) Input layer normalizes input data maps as follows to make VWNN get better converging performance.

    where X is input data map sizing (m,n), μ is a column vector whose row elements are row mean values of X row, and σ is a column vector whose row elements are row standard deviation of X,and no element of σ equals zero. Y is output data map sizing (m,n).

    (2) Activation layer functions as introducing activation function. In this paper, considering the good performance of LeakyReLU, it is used as activation function.16

    where X is input data map sizing (m,n), W is weights matrix sizing (m,n),b is a bias vector sizing (m,1), and y is an output vector sizing (m,1).

    The back propagation of feature separation layer is shown as

    where lyis output loss,lxis input loss,X is the input matrix of feature input layer, DWis gradients of weights, dbis gradients of biases, and η is learning rate.

    (4) Variable-weights layer makes weights a function of input so that the algorithm could get stronger nonlinear expressive capability. Theoretically, due to the fact that variable-weights algorithm is a mapping relationship from input to weights, it can be used in all weightsrelated operation, such as fully connected layer, LSTM and convolution layer.

    Considering that any input array could be transformed into a vector, without the loss of generality, this paper studies the situation where input and output are all vectors. Forward propagation proceeds as the following formulas:

    where x is input vector(m,1),V is variable-weights tensor sizing (m,n,m), W is weights matrix (m,n), B is bias matrix(m,n), o is offset vector (n,1).

    As shown in Eq.(14),W(x)is the mathematical description of the variable-weights thought,where W(weights)is the function of x(input). Eq. (15) is the specific nonlinear relationship from x to W. Instead of updating weights, V is updated by back propagation using Eq. (16) demonstrated below.

    Back propagation is shown as follows. Since weights are decided by variable-weights tensor, W does not need to be updated.

    where DVis gradients of variable-weights tensor, DBis gradients of biases, and dois gradients of offset biases.sum(· , [2,3]) means getting summary along the second and the third dimension of a tensor, which means after sum(· , [2,3]), a tensor will become a vector.

    To make it easier and more friendly to readers, a graphic back propagation of lx(Eq. (22)) is made as an example, as shown in Fig. 2.

    (5) Output layer defines loss function as

    3. Experiments, results and discussion

    3.1. Brief introduction about experiment object

    The basic structure of variable cycle engine and design point’s parameters are the same with Aygun and Turan’s paper.17Yao et al.’s volume dynamics is applied in variable cycle engine’s modeling of this paper.18Limited by space of this paper, a brief introduction to variable cycle engine and data set used for training, validation and test is given.

    Fig.3 is VCE structure diagram.Main components of VCE include Inlet (Inl), Fan (Fan), Core Driven Fan (CDF), High Pressure Compressor (HPC), Combustion (Cbt), High Pressure Turbine (HPT), Low Pressure Turbine (LPT), Mixer(Mix), Nozzle (Noz) and Bypass (Bps). Besides, Mode Switch Valve (MSV), Forward Variable Bypass Ejector (FVBE), and Back Variable Bypass Ejector(BVBE)are used to change flow area.

    Degradation coefficients are defined as

    where Dcom,˙mand Dcom,erepresent component’s degradation coefficients for mass flow and adiabatic efficiency respectively;˙mcom,degradationis a component’s mass flow after degradation and ˙mcom,nominalis nominal mass flow before degradation;ecom,degradationis a component’s adiabatic efficiency after degradation and ecom,nomimalis nominal adiabatic efficiency before degradation.

    In this paper, ten degradation coefficients are defined and act on VCE simultaneously─Dfan,˙m, Dfan,e, Dcdf,˙m,Dcdf,e,Dhpc,˙m,Dhpc,e,Dhpt,˙m, Dhpt,e,Dlpt,˙m, Dlpt,e. Their ranges are all from 0.96 to 1.

    Original data set sizing 27×100000 includes multi-variable dynamic degrading simulation data of VCE. There is one row of time series with interval of 0.02 s, 2 rows of environmental variables (H,Ma), 5 rows of control variables(Wf,A8,Msv,ope,Fvbe,ope,Bvbe,ope) and 19 rows of state variables(T1,P1,T2,P2,T21,P21,T22,P22,T7,P7,T8,P8,P5,T6,Nl,Nh,P3,T5,P6).All variables are listed in Table 1.Rows from 2 to 22 are used as input data.Nl,Nh,P3,T5,P6are variables to be estimated.In this paper, we only take three typical signals’ estimation as example,that is Nl,P3,T5.In order to verify the basic feasibility of the proposed method, the effects of Cdf stator angle,Hpc stator angle, and turbine guide vane angle are all neglected to simplify the problem, which will be researched in further study.

    Input data maps are made into 21 × 9, which includes 21 features (Rows 2-22) and 9 continuous moments (t-8,t).Moment t is the estimated moment.Considering that the problem to be solved is a dynamic problem, more than 1 moments are needed.9 moments are chosen as empirical value,which is a trade-off between inertia of VCE and fluid’s speed. Of course, 7,8,10,13 or other number of moments could be chosen,and they can also achieve similar effects.In machine learning field, parameters like the number of moments are called hyper-parameters, and the choice of hyper-parameters is still an open question. Due to the fact that the choice of hyperparameters is not the focus of this paper, we will not study the empirical value in detail.

    During the process of collecting data,2 environmental variables, 5 control variables and 10 degradation coefficients vary as

    where iuand idare upper bound and lower bound of variables and coefficients,c(t)is a factor ranging from 0 to 1,and o(t)is the value of variables and coefficients at time t.

    c(t)is a periodical function as shown in Eq.(28),including sinusoidal signal, slope signal, and step signal. Besides, in order to make the process more complicated and challenging,periods of control variables (T) are chosen as prime numbers,as shown in Table 2.

    For actual life cycle of aeroengines,degradation coefficients will not change as fast as shown in Table 3,19but considering that there are many uncertainties that cannot be reflected by math model, we decide to make degradation coefficients change more violently to balance those factors. Also, we believe overcoming a harsher situation would make VWNN more convincing and compatible with easier situations.

    Fig. 4 illustrates how to segment original data map sizing 27 × 100000 into input data maps sizing 21 × 9 and corresponding output Nl. tm9is current moment, tm1is initial moment of input data maps, and Nl(23rd row, tm9) is corresponding output. With the window that includes input data maps and corresponding output sliding alongside time direction, 99992 input data maps are made. Actually Nl, P3,T5share the same input data maps. The only difference in the training, validation and test among the three variables is that the output should be altered with another variable, for example, altering Nlwith P3.

    Table 1 Explanation of symbols in original data set.

    For example,21×10000 data map could be segmented into 99992 data maps sizing 21×9,but in order to increase the generalization capability of the proposed method, only 90000 maps are stochastically chosen and used to train, validate and test the proposed method. 70000 input data maps are stochastically chosen and used to train VWNN, 10000 input data maps are stochastically chosen and used to validate,and 10000 input data maps are stochastically chosen and used to test.To be noted,data maps for training,validation and test are continuous respectively, which means, for example, 10000 data maps for test representing 10000 continuous moments.

    3.2. Steady performance of VWNN

    As shown in Table 4 and Table 5,10 steady points are stochastically chosen to validate steady performance of VWNN. Different to Section 3.1 where all control variables,environmental variables and degradation coefficients are always changing,this experiment is aimed at validating steady performance of the proposed algorithm,so the test data maps are not actually segmented from the 21 × 100000 data map mentioned in Section 3.1. Actually, by setting control variables, environmental variables and degradation coefficients as shown in Table 4 and Table 5, status parameters, like pressure and temperature,could be sampled from variable cycle engine mathematical model. After that, by combining control variables, environmental variables and status parameters, one 21 × 1 data map corresponding to one steady point could be made.Finally, in order to adapt to the input size of VWNN, the 21×1 data map is replicated for nine times alongside the time direction so that the steady data map sizing 21 × 9 could be acquired. Ten steady points correspond to ten 21 × 9 steady data maps. Besides, due to the existence of step signals as shown in Eq. (28), original data maps include many steady processes,which is why the proposed method could be applied to steady process.

    In practice,degradation coefficients will follow certain natural law,and control variables are also constrained by control law, which means that there is dependence between variables and degradation coefficients. However, when components fail or meet harsh conditions, the dependence is no longer valid,like accelerating degradation or the surge of compressor. In these situations, statistical distribution of theses variables will change. In another word, statistical distribution of variables and degradation coefficients tend to be independent and stochastic. In Table 4 and Table 5, all variables and degradation coefficients are sampled randomly and subjected to Gauss distribution. Other distributions can also be used, but that is not the focus of this paper.

    Table 2 Periods of environmental variables and control variables.

    Table 3 Periods of degradation coefficients.

    Table 4 Values of environmental variables and control variables of 10 steady points.

    Table 5 Values of degradation coefficients of 10 steady points.

    Table 6 Analytical redundancy of Nl,P3,T5 based on VWNN at 10 steady points.

    AV stands for actual value,and EV is estimation value.RE means relative error,which is defined as Eq(29).ARE is average value of all relative errors. R. denotes results, and D.denotes discussion.

    R.Table 6 demonstrates analytical redundancy of Nl,P3,T5based on VWNN.It can be seen that VWNN performs best in P3estimation with 0.61% ARE, while less accurate in Nland T5with 1.25% and 1.11% ARE respectively.

    D. Under condition that ARE is smaller than 1.5%, it can be assumed that analytical redundancy based on VWNN is effective and satisfactory. The method to further lower ARE could be combining variable-weights algorithm and CNN, by which way network’s nonlinear expressive capability might be stronger. As for why steady performances of analytical redundancy based on VWNN between Nl,P3,T5are different,the most possible reason is that the complexity of mapping relationship from input to Nl,P3,T5is different.

    3.3. Dynamic performance of VWNN

    In this section, we will compare dynamic performance of analytical redundancy based on VWNN,mobile net and dense net with test data set mentioned in Section 3.1. Mobile net is a lightweight neural network owning 153 layers and usually used in lightweight devices like mobile phones and automobiles.20The amount of dense net’s layers reaches up to 709, and its mainly application situation is where higher accuracy is required with less consideration of calculation speed.21Mobile net is typical of the balance between calculation speed and nonlinear expressive capability, while dense net pursuits extraordinary nonlinear expressive capability but basically neglects the effect of large amounts of parameters. In Figs. 5-7, actual value refers to data from test data, and estimation value is analytical redundancy offered by variableweights neural network.10000 test data maps used in this section are data maps segmented from original data map as mentioned in Section 3.1.

    R. Figs. 5-7 are dynamic performance of Nl,P3,T5analytical redundancy based on VWNN,mobile net and dense net.In Fig.5, it can be seen that Nlanalytical redundancies based on mobile net and VWNN are similarly accurate, both of which are more precise than dense net, but Nldynamic curve based on dense net is smoother and less oscillating. Fig. 6 shows the results of P3analytical redundancy. All three nets’ estimation accuracy is in little difference.What Fig.7 demonstrates is dynamic performance of T5analytical redundancy.Obviously,dense net is the most accurate algorithm,followed by VWNN,with mobile net taking the last position.

    R. Generally comparing Figs. 5-7, we can see that VWNN performs best in P3estimation,and demonstrates similar accuracy in other two tasks. Mobile net also performs best in P3estimation, but its accuracy for Nlestimation is obviously higher than that for T5. Although the accuracy of dense net is acceptable in Nlestimation, its performances in other two tasks are relatively better, especially in T5estimation where dynamic performances of VWNN and mobile net are unsatisfactory.

    D. Differences in dynamic performance among Nl,P3,T5based on VWNN, mobile net and dense net could result from the complexity of the mapping relationship from input to output and each net’s nonlinear expressive capability. For example, dense net with the most layers generally performs better than the other nets,which testifies the effectiveness of complex topological structures in nonlinear expressive capability.

    3.4. Comparison of calculation efficiency

    Table 7 Training time, layers and calculation speed of VWNN, mobile net and dense net.

    R.Table 7 compares three nets’training time for 6 epochs with batch size 128─Training time, the number of layers─Layer and calculation speed of every forward propagation─Sc. It can be seen that VWNN is the most compact algorithm with only 5 layers, 6 min for training and 3 ms for each forward propagation. Compared to the most complex net─dense net,three indicators of VWNN are dozens of or even hundreds of times less than that of dense net.

    R. Table 8 demonstrates ARE and calculation efficiency of Nl,P3,T5analytical redundancy based on mobile net,VWNN,and dense net. Calculation efficiency Ecis defined as Eq. (30),which is used as an efficiency indicator to evaluate how much accuracy that an algorithm could provide per millisecond.

    Table 8 ARE and calculation efficiency of Nl,P3,T5 analytical redundancy based on mobile net, VWNN and dense net.

    D.Dense net excels in P3and T5estimation, but is less accurate in Nlanalytical redundancy. Although the ARE of dense net in P3analytical redundancy is only 0.62%,the price is huge amounts of parameters and calculation overhead around 192 ms. Obviously, as far as Ecis concerned, VWNN significantly outperforms mobile net and dense net.

    D.Comprehensively considering the steady performance,dynamic performance and calculation efficiency of VWNN,it can be asserted that variable-weights algorithm plays a pivotal role in nonlinear expressive capability.

    4. Conclusions

    (1) From the perspective of object, variable-weights neural network is proposed to construct variable cycle engine’s analytical redundancy, which is a multi-variable,strongly nonlinear,dynamic,and time-varying problem.

    (2) From the view of methodology, by making weights a function of input, variable-weights algorithm increases model’s nonlinear expressive capability at the same time of decreasing the number of parameters.

    (3) Experiments and results demonstrate that steady and dynamic performances of variable-weights neural network are satisfactory, and compared to other two nets,variable-weights neural network’s calculation efficiency is five times more than that of other methods.

    (4) Variable-weights neural network demonstrates best accuracy in dynamic pressure analytical redundancy with only 0.74%, while the average relative error of dynamic temperature and rotation speed analytical redundancy is around 0.6%higher than that of dynamic pressure analytical redundancy.

    Declaration of Competing Interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgements

    This work was supported by the National Natural Science Foundation of China (Nos. 51576097 and 51976089) and Foundation Strengthening Project of the Military Science and Technology Commission, China (No.2017-JCJQ-ZD-047-21).

    国产成人精品久久二区二区91| 午夜视频精品福利| 一级毛片女人18水好多| 两个人视频免费观看高清| or卡值多少钱| 国产一区二区三区视频了| 亚洲在线自拍视频| 免费人成视频x8x8入口观看| 亚洲自拍偷在线| 欧美高清成人免费视频www| 香蕉国产在线看| 国产成人系列免费观看| 搞女人的毛片| 一级作爱视频免费观看| 免费看十八禁软件| www.999成人在线观看| 黄色女人牲交| 久久久国产欧美日韩av| 亚洲成人久久性| 岛国在线观看网站| 久久精品国产综合久久久| 高清在线国产一区| 女人被狂操c到高潮| 在线观看免费午夜福利视频| 亚洲国产欧美一区二区综合| 精品久久久久久久久久免费视频| 禁无遮挡网站| 男女之事视频高清在线观看| 麻豆国产97在线/欧美 | 又黄又爽又免费观看的视频| 国产精品野战在线观看| 色尼玛亚洲综合影院| 亚洲免费av在线视频| 又大又爽又粗| 久久中文字幕人妻熟女| 麻豆国产97在线/欧美 | 床上黄色一级片| 国产真人三级小视频在线观看| 三级国产精品欧美在线观看 | 可以在线观看毛片的网站| 中文字幕最新亚洲高清| 亚洲国产精品久久男人天堂| 国产成人av教育| 69av精品久久久久久| 国产一区二区在线av高清观看| 国产野战对白在线观看| 欧美精品亚洲一区二区| 国产激情偷乱视频一区二区| 高清在线国产一区| 亚洲第一欧美日韩一区二区三区| 这个男人来自地球电影免费观看| 久久久久久久午夜电影| 香蕉av资源在线| av天堂在线播放| 一边摸一边做爽爽视频免费| 黑人操中国人逼视频| 国产av一区在线观看免费| 可以免费在线观看a视频的电影网站| 18禁裸乳无遮挡免费网站照片| 一个人免费在线观看的高清视频| 一区福利在线观看| 此物有八面人人有两片| 久久久久久免费高清国产稀缺| 岛国在线观看网站| 国产亚洲精品av在线| 国产人伦9x9x在线观看| aaaaa片日本免费| 一级作爱视频免费观看| 久久久久九九精品影院| 亚洲aⅴ乱码一区二区在线播放 | 欧美日韩亚洲国产一区二区在线观看| 别揉我奶头~嗯~啊~动态视频| 国产精品 国内视频| 视频区欧美日本亚洲| 中文字幕人妻丝袜一区二区| 亚洲成人国产一区在线观看| 三级国产精品欧美在线观看 | 日本一区二区免费在线视频| 19禁男女啪啪无遮挡网站| 久久久久久九九精品二区国产 | 亚洲精品久久成人aⅴ小说| 婷婷亚洲欧美| 亚洲精品av麻豆狂野| 国产视频内射| 嫁个100分男人电影在线观看| 亚洲av电影在线进入| 琪琪午夜伦伦电影理论片6080| 波多野结衣高清作品| 亚洲国产精品sss在线观看| 亚洲精品在线观看二区| 伊人久久大香线蕉亚洲五| 亚洲 欧美 日韩 在线 免费| 欧美成人午夜精品| 国产黄片美女视频| 欧美精品亚洲一区二区| 国产精品98久久久久久宅男小说| 欧美性长视频在线观看| 一本综合久久免费| 国产熟女xx| 又黄又粗又硬又大视频| 看免费av毛片| 成人18禁高潮啪啪吃奶动态图| 女人被狂操c到高潮| 久久草成人影院| 老司机福利观看| 99久久综合精品五月天人人| 精品午夜福利视频在线观看一区| 国产亚洲欧美98| 最新在线观看一区二区三区| 首页视频小说图片口味搜索| 在线看三级毛片| 免费在线观看视频国产中文字幕亚洲| 99久久精品国产亚洲精品| 午夜福利视频1000在线观看| 99久久无色码亚洲精品果冻| 亚洲精品久久国产高清桃花| 欧美日韩福利视频一区二区| 无人区码免费观看不卡| 精品久久蜜臀av无| 亚洲欧美精品综合久久99| 非洲黑人性xxxx精品又粗又长| 色精品久久人妻99蜜桃| 亚洲av第一区精品v没综合| 国产亚洲精品久久久久5区| 琪琪午夜伦伦电影理论片6080| 美女高潮喷水抽搐中文字幕| 国产精品一及| 国产免费av片在线观看野外av| 国产精品影院久久| 日本五十路高清| 男人舔奶头视频| 此物有八面人人有两片| 日韩欧美国产一区二区入口| 无遮挡黄片免费观看| 全区人妻精品视频| 日本黄色视频三级网站网址| 亚洲精品一卡2卡三卡4卡5卡| 又黄又粗又硬又大视频| 午夜日韩欧美国产| 久久 成人 亚洲| 丁香六月欧美| 精品国产乱子伦一区二区三区| 精品久久久久久久久久久久久| 欧洲精品卡2卡3卡4卡5卡区| 校园春色视频在线观看| 成年女人毛片免费观看观看9| 又爽又黄无遮挡网站| 日本一二三区视频观看| 国模一区二区三区四区视频 | 亚洲片人在线观看| 一区福利在线观看| 国产精品久久久av美女十八| 国产精品影院久久| 在线视频色国产色| 一a级毛片在线观看| 国产免费av片在线观看野外av| 久久这里只有精品中国| 国产精品久久久久久精品电影| 亚洲国产看品久久| 18禁黄网站禁片免费观看直播| 女人高潮潮喷娇喘18禁视频| av中文乱码字幕在线| 国产男靠女视频免费网站| 成人三级黄色视频| 国产精品香港三级国产av潘金莲| 天天躁狠狠躁夜夜躁狠狠躁| av天堂在线播放| 五月伊人婷婷丁香| 亚洲国产看品久久| 日韩精品免费视频一区二区三区| 国产主播在线观看一区二区| 美女黄网站色视频| 久久香蕉国产精品| 俄罗斯特黄特色一大片| 丝袜美腿诱惑在线| 欧美日韩亚洲综合一区二区三区_| 丝袜美腿诱惑在线| 性色av乱码一区二区三区2| 亚洲aⅴ乱码一区二区在线播放 | 色综合婷婷激情| 亚洲欧美日韩东京热| 日本成人三级电影网站| 久久婷婷成人综合色麻豆| 久久久久亚洲av毛片大全| 97人妻精品一区二区三区麻豆| 五月伊人婷婷丁香| 久久久水蜜桃国产精品网| 日韩精品免费视频一区二区三区| 国产野战对白在线观看| 久久久久久人人人人人| 麻豆成人av在线观看| 小说图片视频综合网站| 在线观看免费视频日本深夜| 麻豆成人av在线观看| 熟女电影av网| 日韩中文字幕欧美一区二区| 又紧又爽又黄一区二区| 久热爱精品视频在线9| 欧美激情久久久久久爽电影| 亚洲av五月六月丁香网| 在线十欧美十亚洲十日本专区| 国产成人精品久久二区二区免费| 精品一区二区三区av网在线观看| 中文在线观看免费www的网站 | 国产精品综合久久久久久久免费| 亚洲专区国产一区二区| 久热爱精品视频在线9| 777久久人妻少妇嫩草av网站| 国产亚洲精品久久久久久毛片| 黄色女人牲交| 一进一出抽搐gif免费好疼| 神马国产精品三级电影在线观看 | 在线观看日韩欧美| 成熟少妇高潮喷水视频| 国产激情欧美一区二区| 老司机在亚洲福利影院| 麻豆久久精品国产亚洲av| 18禁黄网站禁片免费观看直播| 精品乱码久久久久久99久播| 亚洲九九香蕉| 嫁个100分男人电影在线观看| 美女 人体艺术 gogo| 午夜久久久久精精品| 国产精品亚洲美女久久久| 国产亚洲av高清不卡| 老熟妇仑乱视频hdxx| 国产精品香港三级国产av潘金莲| 99热只有精品国产| 国内毛片毛片毛片毛片毛片| 国产亚洲av嫩草精品影院| 老司机午夜福利在线观看视频| 久久久久精品国产欧美久久久| 国产激情偷乱视频一区二区| 亚洲专区国产一区二区| 人人妻人人澡欧美一区二区| 国产欧美日韩一区二区三| 久久中文字幕人妻熟女| bbb黄色大片| 久久久久久亚洲精品国产蜜桃av| 黄色丝袜av网址大全| 亚洲最大成人中文| www.自偷自拍.com| 男人舔女人下体高潮全视频| 国产一区二区三区在线臀色熟女| 成人午夜高清在线视频| 国内精品久久久久久久电影| 亚洲第一欧美日韩一区二区三区| 夜夜躁狠狠躁天天躁| 久久久久国产精品人妻aⅴ院| ponron亚洲| 在线免费观看的www视频| 色播亚洲综合网| 国产69精品久久久久777片 | 精品久久久久久久末码| 国产麻豆成人av免费视频| 日本在线视频免费播放| 午夜日韩欧美国产| 国产精品99久久99久久久不卡| 亚洲成a人片在线一区二区| 一级毛片精品| 一区二区三区国产精品乱码| 国产一区二区三区视频了| 国产三级中文精品| 久久香蕉精品热| 亚洲无线在线观看| 日韩欧美在线二视频| 婷婷精品国产亚洲av| 国产成年人精品一区二区| 欧美黄色淫秽网站| 欧美乱色亚洲激情| 午夜福利免费观看在线| 国产精品野战在线观看| 亚洲美女视频黄频| 中出人妻视频一区二区| 人人妻,人人澡人人爽秒播| 久久精品91无色码中文字幕| 天天一区二区日本电影三级| 麻豆国产97在线/欧美 | 精品国产乱子伦一区二区三区| 精品国产超薄肉色丝袜足j| 嫩草影院精品99| 最近最新中文字幕大全电影3| 日韩中文字幕欧美一区二区| 日本a在线网址| 可以在线观看毛片的网站| 草草在线视频免费看| bbb黄色大片| 欧美日韩精品网址| 久久中文字幕人妻熟女| 久久国产精品人妻蜜桃| 婷婷六月久久综合丁香| 91麻豆av在线| 亚洲自拍偷在线| 别揉我奶头~嗯~啊~动态视频| 不卡av一区二区三区| 88av欧美| 亚洲欧洲精品一区二区精品久久久| 12—13女人毛片做爰片一| 又大又爽又粗| 国产免费男女视频| 听说在线观看完整版免费高清| 久久伊人香网站| 村上凉子中文字幕在线| 精品无人区乱码1区二区| 亚洲国产高清在线一区二区三| 亚洲av日韩精品久久久久久密| 欧美国产日韩亚洲一区| 色噜噜av男人的天堂激情| 99热6这里只有精品| 国产午夜福利久久久久久| 色播亚洲综合网| 国产熟女xx| 日韩三级视频一区二区三区| 国产成人av激情在线播放| 无遮挡黄片免费观看| 毛片女人毛片| 999久久久精品免费观看国产| 少妇裸体淫交视频免费看高清 | 久久精品夜夜夜夜夜久久蜜豆 | 亚洲中文字幕一区二区三区有码在线看 | 成熟少妇高潮喷水视频| 在线播放国产精品三级| 男男h啪啪无遮挡| 久久久久免费精品人妻一区二区| 国产午夜精品论理片| 在线观看免费视频日本深夜| 日韩精品青青久久久久久| 亚洲国产看品久久| 麻豆国产97在线/欧美 | 欧美极品一区二区三区四区| 久久香蕉精品热| 免费高清视频大片| 成人av在线播放网站| 看黄色毛片网站| 久久国产乱子伦精品免费另类| 99精品欧美一区二区三区四区| 久久人妻福利社区极品人妻图片| 午夜精品在线福利| 精品久久久久久久久久久久久| 99久久99久久久精品蜜桃| av免费在线观看网站| 美女免费视频网站| 首页视频小说图片口味搜索| 婷婷精品国产亚洲av在线| 午夜福利视频1000在线观看| 精品不卡国产一区二区三区| 欧美黑人巨大hd| 搞女人的毛片| 日本免费a在线| 久久久久久久久中文| 亚洲av成人av| 高潮久久久久久久久久久不卡| 午夜a级毛片| 精品日产1卡2卡| 亚洲乱码一区二区免费版| 欧美色视频一区免费| 婷婷精品国产亚洲av| 99国产精品一区二区蜜桃av| 国内精品一区二区在线观看| 性色av乱码一区二区三区2| 搡老熟女国产l中国老女人| 90打野战视频偷拍视频| 国产99久久九九免费精品| 久久精品国产清高在天天线| 伦理电影免费视频| 色综合亚洲欧美另类图片| 人妻久久中文字幕网| 午夜免费成人在线视频| 露出奶头的视频| 国产精品自产拍在线观看55亚洲| 亚洲自偷自拍图片 自拍| 黑人欧美特级aaaaaa片| 成年女人毛片免费观看观看9| 99久久综合精品五月天人人| 成熟少妇高潮喷水视频| 99久久99久久久精品蜜桃| 亚洲av成人精品一区久久| 可以在线观看毛片的网站| 国产高清视频在线播放一区| 九色成人免费人妻av| 后天国语完整版免费观看| 精品国内亚洲2022精品成人| 久久精品91蜜桃| 国产一区二区在线观看日韩 | 日本五十路高清| 日韩有码中文字幕| 91字幕亚洲| 免费在线观看黄色视频的| 999精品在线视频| 免费看十八禁软件| 国产亚洲精品久久久久久毛片| av视频在线观看入口| 少妇熟女aⅴ在线视频| 制服丝袜大香蕉在线| 好看av亚洲va欧美ⅴa在| 叶爱在线成人免费视频播放| 亚洲午夜理论影院| 极品教师在线免费播放| 欧美性长视频在线观看| 亚洲国产欧美一区二区综合| 亚洲av五月六月丁香网| 精品久久久久久久末码| 国产精品亚洲一级av第二区| 男插女下体视频免费在线播放| 在线a可以看的网站| 亚洲五月天丁香| 成人高潮视频无遮挡免费网站| 国产精品精品国产色婷婷| 成人永久免费在线观看视频| 51午夜福利影视在线观看| 亚洲午夜理论影院| а√天堂www在线а√下载| 欧美性长视频在线观看| 天天躁夜夜躁狠狠躁躁| 亚洲全国av大片| 五月伊人婷婷丁香| 亚洲精品久久成人aⅴ小说| 午夜福利欧美成人| 亚洲性夜色夜夜综合| 亚洲av第一区精品v没综合| videosex国产| 嫁个100分男人电影在线观看| 校园春色视频在线观看| 777久久人妻少妇嫩草av网站| 午夜福利欧美成人| 亚洲人成77777在线视频| 非洲黑人性xxxx精品又粗又长| 国产激情欧美一区二区| 亚洲乱码一区二区免费版| 母亲3免费完整高清在线观看| 一卡2卡三卡四卡精品乱码亚洲| 国产精品乱码一区二三区的特点| 无遮挡黄片免费观看| АⅤ资源中文在线天堂| 久久天躁狠狠躁夜夜2o2o| 国产精品香港三级国产av潘金莲| 久久精品国产99精品国产亚洲性色| www.精华液| 亚洲国产日韩欧美精品在线观看 | 午夜两性在线视频| 99久久精品热视频| a在线观看视频网站| 国产在线精品亚洲第一网站| 欧美三级亚洲精品| 久久精品综合一区二区三区| 亚洲免费av在线视频| 欧美精品亚洲一区二区| 国产精品九九99| 女警被强在线播放| 婷婷亚洲欧美| 天堂动漫精品| 国产视频内射| 国内精品一区二区在线观看| 99精品欧美一区二区三区四区| 免费在线观看黄色视频的| 国产伦在线观看视频一区| av欧美777| 国产真实乱freesex| 欧美日本视频| 男人舔女人下体高潮全视频| 男女视频在线观看网站免费 | 狠狠狠狠99中文字幕| 在线观看66精品国产| 一本久久中文字幕| 又粗又爽又猛毛片免费看| aaaaa片日本免费| 中文字幕人成人乱码亚洲影| 精品熟女少妇八av免费久了| 欧美乱色亚洲激情| 国产91精品成人一区二区三区| www.999成人在线观看| 久久婷婷成人综合色麻豆| 国产午夜精品久久久久久| 精品午夜福利视频在线观看一区| 欧美不卡视频在线免费观看 | 首页视频小说图片口味搜索| 久久精品91蜜桃| 欧美zozozo另类| 国产精品电影一区二区三区| 国产成人系列免费观看| 少妇裸体淫交视频免费看高清 | 国产精品98久久久久久宅男小说| 久久精品国产亚洲av香蕉五月| 成年人黄色毛片网站| 国产午夜精品论理片| 久久久国产欧美日韩av| 国产成人精品无人区| 国产又黄又爽又无遮挡在线| 好男人电影高清在线观看| 午夜亚洲福利在线播放| 欧美中文综合在线视频| 亚洲精品av麻豆狂野| 妹子高潮喷水视频| or卡值多少钱| 欧美丝袜亚洲另类 | 日韩欧美一区二区三区在线观看| 久久午夜亚洲精品久久| 亚洲一区中文字幕在线| 久久精品国产清高在天天线| 精品无人区乱码1区二区| 亚洲中文字幕日韩| 天天添夜夜摸| 日本三级黄在线观看| 激情在线观看视频在线高清| 午夜福利在线观看吧| 国产99白浆流出| e午夜精品久久久久久久| 国产三级在线视频| 亚洲国产精品久久男人天堂| 久久香蕉国产精品| 激情在线观看视频在线高清| 麻豆国产av国片精品| 少妇人妻一区二区三区视频| 亚洲欧洲精品一区二区精品久久久| 岛国视频午夜一区免费看| 男男h啪啪无遮挡| 欧美在线一区亚洲| 一边摸一边做爽爽视频免费| 麻豆成人午夜福利视频| 亚洲国产精品sss在线观看| 亚洲欧洲精品一区二区精品久久久| 国产欧美日韩精品亚洲av| 老司机深夜福利视频在线观看| 色播亚洲综合网| 日韩欧美 国产精品| 首页视频小说图片口味搜索| 久久久久久久久久黄片| 国产亚洲精品av在线| 在线观看午夜福利视频| www.自偷自拍.com| 精品欧美一区二区三区在线| 精品乱码久久久久久99久播| 午夜日韩欧美国产| 久久亚洲真实| 嫩草影视91久久| 欧美中文日本在线观看视频| 深夜精品福利| 91九色精品人成在线观看| a级毛片在线看网站| 国产精品综合久久久久久久免费| av片东京热男人的天堂| 少妇被粗大的猛进出69影院| 18禁裸乳无遮挡免费网站照片| 亚洲性夜色夜夜综合| 成人国语在线视频| 久久国产精品人妻蜜桃| 久久国产精品影院| 久久国产乱子伦精品免费另类| 成人国产一区最新在线观看| 国产精品久久久av美女十八| 女人高潮潮喷娇喘18禁视频| 在线免费观看的www视频| 亚洲色图 男人天堂 中文字幕| 亚洲国产中文字幕在线视频| 成人永久免费在线观看视频| 中文字幕熟女人妻在线| 高潮久久久久久久久久久不卡| 欧美在线黄色| 久久久精品国产亚洲av高清涩受| 一区二区三区激情视频| av视频在线观看入口| 9191精品国产免费久久| 亚洲av美国av| 脱女人内裤的视频| 亚洲精品美女久久av网站| 美女扒开内裤让男人捅视频| 午夜福利18| 成人国语在线视频| av天堂在线播放| av片东京热男人的天堂| 亚洲精品在线观看二区| 午夜激情福利司机影院| 午夜视频精品福利| 久久久精品国产亚洲av高清涩受| 少妇粗大呻吟视频| 久久久精品国产亚洲av高清涩受| 午夜视频精品福利| 无人区码免费观看不卡| 国产av在哪里看| 亚洲第一欧美日韩一区二区三区| 久久久久久人人人人人| 午夜福利视频1000在线观看| 久久精品91蜜桃| 国产真实乱freesex| 男人舔女人的私密视频| 1024手机看黄色片| 亚洲专区中文字幕在线| 日韩有码中文字幕| 又粗又爽又猛毛片免费看| av片东京热男人的天堂| 日韩精品青青久久久久久| 在线看三级毛片| 欧美日本亚洲视频在线播放| 非洲黑人性xxxx精品又粗又长| 久久久久久人人人人人| 成人av在线播放网站| 男人舔女人的私密视频| 国产精品美女特级片免费视频播放器 | 久久精品综合一区二区三区| 成人精品一区二区免费| 精品久久久久久久毛片微露脸| 窝窝影院91人妻| 久久久久久国产a免费观看| 在线观看舔阴道视频| 丝袜美腿诱惑在线| 91九色精品人成在线观看| 国产伦人伦偷精品视频| 久久精品国产综合久久久| 露出奶头的视频| 少妇熟女aⅴ在线视频| 中国美女看黄片| 久久精品91蜜桃| 亚洲成人久久爱视频| 2021天堂中文幕一二区在线观| 在线看三级毛片|