• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Predicting solutions of the Lotka-Volterra equation using hybrid deep network

    2022-03-04 09:56:54ZiFeiLinYnMingLingJiLiZhoJioRuiLi

    Zi-Fei Lin, Yn-Ming Ling, Ji-Li Zho, Jio-Rui Li

    a School of Statistics, Xi’an University of Finance and Economics, Xi’an 710100, China

    b China (Xi’an) Institute for Silk Road Research, Xi’an 710100, China

    Keywords:Lotka-Volterra equations HDN-LV algorithm Stochastic Lotka-Volterra equations Parameter optimization

    ABSTRACT Prediction of Lotka-Volterra equations has always been a complex problem due to their dynamic properties. In this paper, we present an algorithm for predicting the Lotka-Volterra equation and investigate the prediction for both the original system and the system driven by noise. This demonstrates that deep learning can be applied in dynamics of population. This is the first study that uses deep learning algorithms to predict Lotka-Volterra equations. Several numerical examples are presented to illustrate the performances of the proposed algorithm, including Predator nonlinear breeding and prey competition systems, one prey and two predator competition systems, and their respective systems. All the results suggest that the proposed algorithm is feasible and effective for predicting Lotka-Volterra equations. Furthermore, the influence of the optimizer on the algorithm is discussed in detail. These results indicate that the performance of the machine learning technique can be improved by constructing the neural networks appropriately.

    Recently, big data has developed rapidly, and data-based deep learning algorithms have been widely used, from text translation and information recognition to predicting stock prices and disease spread. The results predicted by this method are obtained by using a neural network algorithm with multiple hidden neural states [1].These hidden neural states form a "black box" model with trained internal parameters. We only observe the final model output when given a set of measured training data [2]. For the parameters in the model, time series data is usually used for training, which means that the prediction object of the deep learning model is usually time series data. Time series data in real situations, such as financial data, environmental data, meteorological data, etc., are chaotic. In the mathematical description, we call it chaotic motion.A chaotic time series is generated by the deterministic dynamics of a nonlinear system, which is sensitive to initial conditions and is challenging to predict for such systems. Most scholars use mathematical methods to predict it; these methods can predict accurately in theory. However, they cannot consider future development and change laws and development levels of market phenomena when they make actual predictions. There will be a large deviation when there is a large change in the outside world. Therefore,traditional statistical techniques cannot give reliable predictions for nonlinear dynamical systems.

    Machine learning has good results in the prediction of dynamic systems. Scholars have carried out research in this field. Combining machine learning with knowledge-based modeling, LeCun et al.[1] predicted predictions for large, complex spatiotemporal systems. Testing with the Kuramoto–Sivashinsky equation confirms that the proposed method not only can be scaled to provide excellent performance for very large systems, but the time-series data length required to train multiple parallel machine learning components is significantly lower than without parallelization desired length. Wang et al. [3] proposed to predict chaotic dynamics using a neural network with an inhibitory mechanism, and the accuracy of the proposed algorithm was confirmed by simulating a well-known nonlinear dynamical system prototype. And it has been proved that the algorithm is also suitable for predicting the partial differential system. Pyragas and Pyragas [4] used reservoir computing algorithms to predict and prevent extreme events.They used FitzHugh-Nagumo systems as an example to demonstrate the feasibility of the proposed method, with excellent results. Pathak et al. [5] also used the reservoir calculation algorithm to predict the chaotic dynamic system. They confirmed that the proposed method is accurate in short-term predictions by the Lorenz system and Kuramoto-Sivashinsky equations. At the same time, the predicted results can perfectly restore the dynamic characteristics of the original system. Vlachas et al. [6] propose complex spatiotemporal dynamics prediction using a backpropagation algorithm and reservoir computing in recurrent neural networks.They found that when the whole state dynamics are available for training, the RC algorithm outperforms the BPTT method in predicting performance and capturing long-term statistics while requiring less training time. Nevertheless, large-scale RC models can be unstable and more likely to diverge than BPTT algorithms in the case of reduced-order data. In contrast, the neural network trained by BPTT has superior predictive ability. It can well capture the conclusion of the dynamics of low-order systems, which once again confirms the applicability of deep learning in predicting chaotic systems. Cui et al. [7] applied complex network theory to construct echo-state network dynamic reservoirs. They proposed three echo-state networks with different dynamic reservoir topologies. They conducted simulation experiments and detailed analysis and found that the proposed echo state network has higher prediction accuracy and enhanced echo characteristics but still has the same short-term storage capacity. Falahian et al. [8] proposed using a multilayer feedforward neural network to predict some wellknown chaotic systems while evaluating the accuracy and robustness of the model, and the results proved reliable. Besides, they utilize the proposed neural network to represent the modeling results of this chaotic response model. The ability of the neural network to simulate this specific brain response was demonstrated.Follmann and Rosa [9] used reservoir computational algorithms to predict the temporal evolution of neuronal activity resulting from the Hindmarsh-Rose neuron model. Accurate short- and longterm predictions of cyclical (tonic and explosive) neuronal behavior were obtained, but only accurate short-term predictions of chaotic neuronal states. However, the studies mentioned above all stay in biology and meteorology but do not involve population dynamics.Population dynamics are fundamental to our lives.No population in nature is isolated; any complex ecosystem is a complex network formed by the interaction of multiple populations, such as intraspecies competition, cooperation, interspecies predation, and predation.

    This paper proposes a deep hybrid neural network for predatorprey model prediction. We predict nonlinear predator reproduction and prey competition systems, predator nonlinear predator reproduction and prey competition stochastic systems, and one-prey and two-predator competition systems, respectively, and find that the proposed algorithm, compared with traditional neural network algorithms, can Improve prediction accuracy. At the same time, for the dynamic characteristics of the system, the results predicted by the algorithm can also be perfectly reproduced. In addition, we also discuss the algorithm’s more interesting issue: the optimizer’s choice.

    The Lotka-Volterra equation is a dynamical system consisting of two first-order nonlinear differential equations, also called the predator-prey equation, proposed by Lotka and Volterra in 1296[10–12]. It is often used to describe the dynamics of predators in biological systems as they interact with their prey. Specifically,the size of the two populations waxes and wanes. The dynamical model that can describe the size of the two populations was initially proposed as

    Fig. 1. Convolutional neural network.

    In Eq. (2), all terms are replaced by the corresponding population density functions. WhereB(x,y)represents the rate at which the prey is consumed, i.e., the predation rate, it is a monotonic non-decreasing function that depends on the population density of predators and prey.D(x,y)is called the fertility function and represents the reproductive rate of the predator. In the original expression, we considered that the relationship between the biomass of the prey and the biomass of the predator is constant; keeping this assumption and considering other additional factors, the formD(x,y)is formed.A(x)represents the reproduction of the prey itself, andC(y)represents the reproduction of the predator itself.

    Convolutional neural networks are a class of feedforward neural networks that include convolutional computation and have a deep structure, and are one of the representative algorithms of deep learning. Research on it began in the 1980s and 1990s, with timedelay networks and LeNet-5 being the first convolutional neural networks to appear [13]. After the twenty-first century, with the proposed deep learning theory and the improvement of numerical computing devices, convolutional neural networks have been developed rapidly and applied to computer vision, natural language processing, and other fields [14]. It differs from an ordinary neural network by including a feature extractor consisting of a convolutional layer and a subsampling layer (pooling layer). In the convolutional layer of a convolutional neural network, a neuron is connected to only some of the neighboring neurons and usually contains several feature maps; each feature map consists of several rectangularly arranged neurons, and the neurons of the same feature map share weights, where the shared weights are the convolutional kernel. For subsampling (pooling), there are usually two forms of mean pooling and max pooling, which can be regarded as a unique convolution process. Convolution and subsampling greatly simplify model complexity and reduce model parameters. In order to get the output feature mapY, we use the convolution kernelsW1,W2,...,Wpto convolve the input feature mapsX1,X2,...,Xnrespectively, then add the convolution results and add the upper bias termb, the net inputzof the convolution layer is obtained, and the output feature mapYis obtained by performing a nonlinear transformation on it. After the convolutional layer operation is performed, the obtained feature mapYis used for feature selection to reduce the number of features, thereby reducing the number of parameters and preventing overfitting. Figure 1 shows the exact mechanism of operation of the algorithm.

    Fig. 2. Gated recurrent units neural network.

    The traditional recurrent neural network has some shortcomings that are difficult to overcome. One of the biggest problems is gradient disappearance and gradient explosion [15–17]. To solve these problems and improve prediction accuracy, Cho [18] proposed a gated recurrent unit neural network (GRU) in 2014. GRU proposes using two vectors, the update gate. It resets the gate to decide what information should be passed to the output. It can save information from a long time ago and remove irrelevant information, enhancing the neural network’s memory capacity and making the final classification and regression more accurate. The update gate controls how much state information from the previous moment is brought into the current state. The larger the value,the more state information from the previous moment is brought in; the reset gate controls how much state information from the previous moment is ignored. The smaller the value, the more it is ignored. From Fig. 2,Z(t)represents the update gate, which acts similarly to the forgetting and input gates in the LSTM algorithm,deciding what information to discard and what new information to add.Rtis the reset gate, the reset gate is used to decide the degree of discarding the previous information, andH(v)is the final predicted time series data result. Equations (3)–(6) illustrates how the GRU are calculated, whereWrepresents the weight.

    In cognitive science, due to bottlenecks in information processing, humans selectively focus on a portion of all information while ignoring other visible information. The above mechanism is often referred to as the attention mechanism. For a classification task,we extract image features and classify the images into different classes based on the differences between these features. However,only useful features contribute to the classification, while different features have different meanings [19–21]. The method has been widely used in text recognition and image recognition. It has been confirmed that it is effective in extracting features.

    Figure 3 is the input data, which is passed to the softmax layer by Eq. (7) to obtain the attention distribution, let it be, which can be interpreted as the degree of attention of the message. Then pass through Eq. (8) to finally get the required informationa.We introduced some basic algorithms of neural networks, which can obtain prediction results but are limited by their respective drawbacks, and the accuracy of the obtained prediction results is often poor. We will describe the numerical scheme for predicting the Lotka-Volterra equation by deep neural networks and name it the HDN-LV algorithm.

    Fig. 3. Attentional mechanism.

    We propose a form of neural network connectivity. We use a layer of two-dimensional convolutional neural network and a layer of attention mechanism for feature extraction of the features present in the original data. We finally pass them to the traditional DNN fully connected layer to produce the final output. Next, the GRU model trains and predicts the output data to achieve better prediction results. In addition, we introduce a regularization factor as a penalty term in the algorithm to achieve higher prediction accuracy.

    We will construct the HDN-LV algorithm. For the selection of the prediction range, the range of the independent variables of the Lotka-Volterra equation needs to be truncated since the prediction range of the deep hybrid neural network should be a finite set.Considering that in practice, the usual forecasts often do not require particularly long periods, we should choose a suitable interval. According to the literature [22], it is known that the currently obtained prediction horizon is about 5 to 6 Lyapunov times, where one Lyapunov time is the reciprocal of the maximum Lyapunov exponent. So we calculate the Lyapunov exponent for the system’s initial solution to determine the prediction’s final interval.

    Algorithm 1 is the implemental process of the proposed HDNLV algorithm, and four different systems will be given to test the validity of our method in the next step.

    Then we select predator nonlinear reproduction prey competition systems, one prey and two predator competition systems, and stochastic systems with noise added to confirm that HDN-LV methods can predict the dynamical evolution of ecosystems.

    Algorithm 1 HDN-LV algorithm.

    Fig. 4. Numerical solution of predator nonlinear reproduction and prey competition system.

    Based on the numerical solution, we calculate the largest Lyapunov exponent for Eq. (10) as 0.1771. Therefore, the prediction time is chosen as [0,200]. We set the penalty factor of the algorithm to 6×10?4. The size of the network is set to four hidden layers with 80 nodes in each hidden layer. Meanwhile, the Adam optimization technique is applied, getting the minimum value of the loss function is about 10?6. Figure 5 and Table 1 show the obtained results. The blue dashed line is the HDN-LV solution, and the 10–6green dashed line is the GRU solution. Compared to the conventional GRU algorithm, the solution of the HDN-LV algorithm is almost identical to the numerical solution marked by the solid red line, with an average accuracy of 99.98%. These results show that the HDN-LV method effectively predicts the Lotka-Volterra equation.

    We investigated the predicted dynamical characteristics of predator nonlinear reproduction and prey competition systems,and the results were ideal. However, the study of this situation alone does not exactly match the reality. In real life, the system may also be affected by factors other than predators and prey, such as weather changes, earthquakes, tsunamis, floods, other natural disturbances, and human interference. In this part, we add random noise to Eq. (10) to confirm the efficiency of the HDN-LV algorithm.We establish the Ito equation for Eq. (14) [23].

    whereσu1,σv1is the noise intensity anddwu,dwvis the standard Wiener process, setσu1=σv1=0.1. The training data of the algorithm is obtained by iterating on Eq. (14) using the fourth-order Runge-Kutta method with the initial values as shown in Fig. 6.

    The prediction interval is also determined by calculating the largest Lyapunov exponent. According to the calculation results,the final prediction interval is [0,180]. Using the HDNN-LV algorithm, the activation function is selected as thetanh function, the loss function is minimized using the Adam method, and the minimum loss value is about 10?6. Where, we set the penalty factor of the algorithm to 1×10?3. Meanwhile, the conventional GRU algorithm is designed to be compared with the HDN-LV algorithm.From Fig. 7a and 7b, it can be seen that the results obtained bythe HDN-LV algorithm prediction are closer to the stochastic solution compared to the traditional GRU algorithm. Meanwhile, from Fig. 7c and 7d, we can see that the absolute error of the HDN-LV algorithm is closer to zero than the GRU algorithm’s absolute error. To confirm the prediction, we calculated the largest Lyapunov exponent. FromTable 2, it can be seen that the largest Lyapunov exponent calculated by the prediction results of the HDN-LV algorithm is closer to the largest Lyapunov exponent calculated by the stochastic solution. This illustrates that the results predicted by the HDN-LV algorithm fit the stochastic solution more closely,both numerically and in terms of kinetic characteristics, which also shows that the HDN-LV algorithm is more effective in predicting the Lotka-Volterra equation.

    Table 1 The accuracy of the HDN-LV method for Eq. (10).

    Fig. 5. The comparison results between the solution obtained from the HDN-LV algorithm, GRU algorithm, and the exact solution. The blue dashed line is the HDN-LV solution, the green dashed line is the GRU solution, and the solid red line is the exact solution.

    Fig. 6. Stochastic solutions of predator nonlinear reproduction competes with prey.

    Fig. 7. (a) and (b): The comparison results between the solution obtained from the HDN-LV algorithm, GRU algorithm, and the stochastic solution. The blue dashed line is the HDN-LV solution, the green dashed line is the GRU solution, and the solid red line is the stochastic solution. (c): Absolute error of HDN-LV algorithm and GRU algorithm. Green represents the absolute error of the HDN-LV algorithm, and yellow represents the absolute error of the GRU algorithm.

    Table 2 Largest Lyapunov exponent.

    Fig. 8. Numerical solution phase diagram of a competitive system of one prey and two predators: (a) Prey-predator1 projection; (b) Prey-predator2 projection; (c)Predator1-predator2 projection.

    Fig. 9. (a) and (c): The comparison results between the solution obtained from the HDN-LV algorithm, GRU algorithm, and the numerical solution. The blue dashed line is the HDN-LV solution, the green dashed line is the GRU solution, and the solid red line is the numerical solution. (d): Absolute error of HDN-LV algorithm and GRU algorithm. Green represents the absolute error of the HDN-LV algorithm, and yellow represents the absolute error of the GRU algorithm.

    Fig. 10. Stochastic solution of the competitive system of one prey and two predators.

    For the initial value of Eq. (16). Setu=1,v1=1,v2=1,γ1=10,γ2=10,α=0.31,β=0.29, assuming the same initial density of predator and prey, and iterate using the fourth-order Runge-Kutta method. Thus determine what happens to the system after it evolves.

    From Fig. 8, the population density of predator1 decreases dramatically after a short period and eventually remains flat around zero. This is due to the competition this predator received from another predator and its influence. At the same time, Predator2 showed a wide range of oscillations in the initial period untilt=6.5, when the density of this population stabilized, fluctuating around 0.8. This indicates that after the initial experience of competition between predator populations, the population is more adapted to the current environment compared to the other predators and reaches equilibrium after getting some depletion. For prey,the same large oscillation experienced in the initial period was caused by competition between the two predator populations. The eventual increase in prey due to the decrease in predator1 led to the eventual stabilization of population density fluctuating around 1.8after the initial oscillation. We use the proposed HDN-LV algorithm for prediction.

    We set the prediction interval to be [0,200]. Using the Adam optimization algorithm, the loss function is minimized by subiterations. The results are shown in Fig. 9a–9c. Here, we set the penalty factor of the algorithm to 5×10?6. The HDN-LV method is used to solve Eq. (16) by setting the network with four hidden layers and 50 nodes in each hidden layer. The maximum absolute error between the HDN-LV and the exact solutions is 10–3.The prediction effect can be seen more intuitively in Fig. 9. Compared with the other traditional algorithm, the prediction effect of the HDN-LV algorithm is better. Its prediction accuracy is 99.91%, the highest among the three methods.

    Fig. 11. (a) and (c): The comparison results between the solution obtained from the HDN-LV algorithm, GRU algorithm, and the numerical solution. The solid blue line is the HDN-LV solution, the green dashed line is the GRU solution, and the solid red line is the numerical solution. (d)-(f): Absolute error of HDN-LV algorithm and GRU algorithm. Blue represents the absolute error of the HDN-LV algorithm, and the green represents the absolute error of the GRU algorithm.

    Table 3 The comparison of Adam and Adam-decay optimizer in the 2D case.

    Table 4 The comparison of Adam and Adam-decay optimizer in the 3D case.

    Meanwhile, the training time is 1651 s when using the HDN-LV method. In comparison, the conventional GRU algorithm takes the 2540 s, and the accuracy of our proposed method is higher than that of the conventional method. This shows that the initial features we designed to extract the data are practical to obtain more accurate prediction results and improve computational efficiency.

    We add random noise to the 3D Lotka-Volterra equation to determine the proposed method’s validity. For Eq. (16), establish the corresponding Ito equation to obtain Eq. (17) [22].

    whereσu,σv1,σv2is the noise intensity anddwu,dwv1,dwv2is the standard Wiener process. Setσu=σv1=σv2=0.1 and iterate on Eq. (17) using the fourth-order Runge-Kutta method to obtain the training data for the algorithm with the initial values as shown in Fig. 10.

    Using the HDN-LV method, we predicted Eq. (17). The predicted interval is [0,190]. For step 1, we choose a training set size of 2000.In step 2, we consider a network with eight hidden layers and 40 nodes in each hidden layer and set the penalty factor to 8×10?4.The weights of the initial algorithm and the deviations are optimized using the Adam optimization algorithm, and the results as shown.

    From Fig. 11a–11c, it can be found that the results obtained by the HDN-LV algorithm are closer to the initial value compared to the other traditional algorithm, where the training loss is about 10?7. The absolute error of the system is shown in Fig. 11d–11f,which clearly shows that the calculation results obtained by the proposed method are closer to zero compared with other algorithms, which is also a side indication of the accuracy and effectiveness of the algorithm.

    In this paper, we propose the HDN-LV algorithm and confirm that it can obtain high accuracy by predicting the Lotka-Volterra equation and the stochastic Lotka-Volterra equation. In this part,we discuss the questions about it: how the optimizer is selected to be used to train the HDN-LV algorithm.

    We optimize the initial algorithm using the Adam optimizer and find that it achieves satisfactory training results only when iterations reach 2×104, which can lead to long training times and cannot get the desired accurate answers fast enough. Therefore, we tried to improve the optimizer. We found it very effective to continue adding decay rate metrics to the Adam optimizer instead of using the adaptive form. We compared the prediction results of the conventional Adam and Adam methods by adding the decay rate index in the two-dimensional and three-dimensional cases. The results are shown in Tables 3 and 4.

    We set the same prediction interval as the traditional method and the parameters of the initial algorithm and use the same computer for its calculation. It can be found that the Adam algorithm with an increased decay rate significantly improved computational efficiency, ensuring that the computation can be completed within 1000 s. In addition, the accuracy of Adam’s method with an increased decay rate is higher than that of traditional optimization methods. Therefore, artificially increasing the decay rate for the Adam method can achieve the same or even better results than the traditional Adam method and save time.

    This paper proposes an HDN-LV method for predicting the Lotka-Volterra equation using deep learning methods. Two examples of 2D and 3D Lotka-Volterra equations and their stochastic forms show that the proposed HDN-LV algorithm has high accuracy. Compared with traditional methods, the method proposed in this paper does not require a deep mathematical foundation. It does not require any transformation of the equations. At the same time, the short computation time and high accuracy are also advantages. Meanwhile, we introduce a regularization factor as the penalty term of the neural network to avoid the initial algorithm overfitting affecting the prediction accuracy.

    In addition, we discuss the selection of the optimizer and the setting of the penalty factor in the HDN-LV method. For the selection of the optimizer, we found that for the adaptive Adam optimizer, setting the decay rate index for it can get the desired prediction results more quickly under the condition of guaranteeing the prediction accuracy.

    It should be noted that while the HDN-LV method is effective for predicting Lotka-Volterra equations driven by white noise,the prediction effectiveness of using this method for other cases,such as color noise-driven Lotka-Volterra equations and broadband and narrowband noise-driven Lotka-Volterra equations, is unknown. Therefore, higher performance and better storage capacity are required to improve the prediction accuracy when solving dynamic systems driven by color noise and wide-band and narrowband noise.

    Declaration of Competing Interest

    The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    Acknowledgment

    The research was supported by the National Natural Science Foundation of China (No. 11902234), Natural Science Basic Research Program of Shaanxi (No. 2020JQ-853), China (Xi’an) Silk Road Research Institute Scientific Research Project (No. 2019ZD02),Shaanxi Provincial Department of Education Youth Innovation Team Scientific Research Project (No. 22JP025), and the Young Talents Development Support Program of Xi’an University of Finance and Economics.

    欧美日韩亚洲高清精品| 日本撒尿小便嘘嘘汇集6| av超薄肉色丝袜交足视频| 又紧又爽又黄一区二区| 黄色怎么调成土黄色| 菩萨蛮人人尽说江南好唐韦庄| 亚洲av国产av综合av卡| 欧美xxⅹ黑人| 香蕉丝袜av| 日本a在线网址| av有码第一页| 中文字幕人妻丝袜一区二区| 啦啦啦 在线观看视频| 国产精品.久久久| 精品久久久久久久毛片微露脸 | 免费高清在线观看日韩| 亚洲欧美一区二区三区黑人| 十分钟在线观看高清视频www| 美女福利国产在线| 亚洲精品一区蜜桃| 亚洲性夜色夜夜综合| 18禁黄网站禁片午夜丰满| 精品国产乱子伦一区二区三区 | 99久久国产精品久久久| 丝袜美足系列| 国产日韩欧美亚洲二区| 日本a在线网址| 爱豆传媒免费全集在线观看| 91九色精品人成在线观看| 午夜视频精品福利| 在线天堂中文资源库| 免费高清在线观看视频在线观看| 国产精品一区二区免费欧美 | 欧美日韩福利视频一区二区| 久久 成人 亚洲| 亚洲国产欧美日韩在线播放| 国产精品二区激情视频| 免费人妻精品一区二区三区视频| 正在播放国产对白刺激| 美女国产高潮福利片在线看| 成年动漫av网址| 最近最新中文字幕大全免费视频| 亚洲综合色网址| 一本综合久久免费| 午夜福利视频在线观看免费| 欧美激情高清一区二区三区| 成人国语在线视频| 无限看片的www在线观看| 波多野结衣av一区二区av| 老司机影院毛片| 久久久精品国产亚洲av高清涩受| 国产免费一区二区三区四区乱码| 韩国精品一区二区三区| 国产一卡二卡三卡精品| 久久精品亚洲av国产电影网| 精品国内亚洲2022精品成人 | 国产免费视频播放在线视频| 精品人妻一区二区三区麻豆| 精品免费久久久久久久清纯 | 操出白浆在线播放| 国产免费一区二区三区四区乱码| 欧美精品高潮呻吟av久久| 久久人妻福利社区极品人妻图片| 成人国语在线视频| 黄色视频在线播放观看不卡| 1024香蕉在线观看| 久久久国产成人免费| 亚洲精品久久午夜乱码| 黑人巨大精品欧美一区二区蜜桃| 老司机影院毛片| 女人精品久久久久毛片| 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲精品自拍成人| svipshipincom国产片| 亚洲精品第二区| a级片在线免费高清观看视频| 国产黄频视频在线观看| 真人做人爱边吃奶动态| 波多野结衣av一区二区av| 成在线人永久免费视频| 在线av久久热| 一二三四在线观看免费中文在| 欧美午夜高清在线| 国产熟女午夜一区二区三区| 亚洲av成人一区二区三| 国产日韩欧美在线精品| 亚洲精品自拍成人| 人人妻人人添人人爽欧美一区卜| 午夜免费成人在线视频| 亚洲一码二码三码区别大吗| 国产日韩一区二区三区精品不卡| 老熟妇仑乱视频hdxx| 国产在线免费精品| 一区二区av电影网| 免费日韩欧美在线观看| 国产精品国产av在线观看| 欧美日韩成人在线一区二区| 亚洲精品久久午夜乱码| 无遮挡黄片免费观看| 丝袜美腿诱惑在线| 国产成人啪精品午夜网站| 97精品久久久久久久久久精品| 黄片小视频在线播放| tocl精华| 在线观看舔阴道视频| 亚洲欧美色中文字幕在线| 在线看a的网站| 法律面前人人平等表现在哪些方面 | 国产在线免费精品| 少妇的丰满在线观看| 在线 av 中文字幕| 另类亚洲欧美激情| 美女主播在线视频| 两性午夜刺激爽爽歪歪视频在线观看 | av片东京热男人的天堂| av免费在线观看网站| 18禁国产床啪视频网站| 欧美日韩一级在线毛片| 欧美 亚洲 国产 日韩一| 一区福利在线观看| 国产熟女午夜一区二区三区| 大片免费播放器 马上看| 国产老妇伦熟女老妇高清| 国产97色在线日韩免费| 午夜福利免费观看在线| 久久人妻福利社区极品人妻图片| 国产一区二区三区av在线| 女人精品久久久久毛片| 久久影院123| av在线播放精品| av在线播放精品| a 毛片基地| 亚洲精品国产区一区二| 国产淫语在线视频| 成人亚洲精品一区在线观看| 久9热在线精品视频| 狂野欧美激情性xxxx| 一本综合久久免费| 啦啦啦啦在线视频资源| 天天影视国产精品| 19禁男女啪啪无遮挡网站| 午夜福利在线观看吧| 国产国语露脸激情在线看| 黑人猛操日本美女一级片| 国产av国产精品国产| av免费在线观看网站| 国产日韩欧美在线精品| 亚洲男人天堂网一区| 纵有疾风起免费观看全集完整版| 黄片大片在线免费观看| 91麻豆av在线| 国产成+人综合+亚洲专区| 午夜视频精品福利| 亚洲,欧美精品.| 亚洲精品国产精品久久久不卡| 色精品久久人妻99蜜桃| 精品欧美一区二区三区在线| 91精品国产国语对白视频| 国产精品免费视频内射| 嫩草影视91久久| 精品人妻在线不人妻| 亚洲性夜色夜夜综合| 99精品久久久久人妻精品| 亚洲avbb在线观看| 日本av手机在线免费观看| 黄色视频在线播放观看不卡| 女人高潮潮喷娇喘18禁视频| 欧美黄色片欧美黄色片| 免费观看av网站的网址| 免费看十八禁软件| 国产国语露脸激情在线看| 国产一区二区三区av在线| 亚洲国产av新网站| 丁香六月欧美| 国产精品熟女久久久久浪| 丝袜脚勾引网站| 国产成人系列免费观看| 国产av一区二区精品久久| 久久国产精品大桥未久av| 一级片免费观看大全| 亚洲人成电影免费在线| 国产精品一区二区在线不卡| 在线观看免费高清a一片| 热99国产精品久久久久久7| 另类精品久久| 日韩电影二区| 亚洲中文av在线| 大型av网站在线播放| 久久久久久亚洲精品国产蜜桃av| 天堂8中文在线网| 9191精品国产免费久久| 国产高清视频在线播放一区 | 男女之事视频高清在线观看| 亚洲色图 男人天堂 中文字幕| 正在播放国产对白刺激| 午夜日韩欧美国产| 97精品久久久久久久久久精品| 午夜免费观看性视频| 人成视频在线观看免费观看| 亚洲人成电影免费在线| 中文字幕av电影在线播放| 国产精品99久久99久久久不卡| 18禁观看日本| www.av在线官网国产| 国产国语露脸激情在线看| 91大片在线观看| 中文字幕精品免费在线观看视频| 日本欧美视频一区| 91成年电影在线观看| 国产成人一区二区三区免费视频网站| 久久青草综合色| 黄色视频在线播放观看不卡| 一区二区日韩欧美中文字幕| 精品免费久久久久久久清纯 | 嫩草影视91久久| 蜜桃在线观看..| 亚洲av日韩在线播放| 人妻久久中文字幕网| 在线观看舔阴道视频| 亚洲伊人色综图| 80岁老熟妇乱子伦牲交| 国产精品.久久久| 99久久国产精品久久久| 啦啦啦在线免费观看视频4| 丝袜脚勾引网站| 亚洲成人国产一区在线观看| a级毛片在线看网站| 午夜成年电影在线免费观看| 久久精品亚洲熟妇少妇任你| 欧美 日韩 精品 国产| 91精品三级在线观看| 免费观看a级毛片全部| 国产精品 欧美亚洲| 午夜两性在线视频| 久久毛片免费看一区二区三区| 91成年电影在线观看| 日本vs欧美在线观看视频| 亚洲精品av麻豆狂野| 国产精品99久久99久久久不卡| 91麻豆精品激情在线观看国产 | 国产av又大| 91大片在线观看| 成人av一区二区三区在线看 | 成人18禁高潮啪啪吃奶动态图| 亚洲五月色婷婷综合| 国产欧美日韩精品亚洲av| 精品熟女少妇八av免费久了| 淫妇啪啪啪对白视频 | 在线精品无人区一区二区三| 久久久久国内视频| 精品久久久久久电影网| 国产一区二区 视频在线| 午夜激情av网站| 在线精品无人区一区二区三| 国产极品粉嫩免费观看在线| 男男h啪啪无遮挡| 国产欧美日韩一区二区三区在线| 考比视频在线观看| 岛国毛片在线播放| 亚洲一区二区三区欧美精品| 首页视频小说图片口味搜索| 脱女人内裤的视频| 两性午夜刺激爽爽歪歪视频在线观看 | 一区二区日韩欧美中文字幕| 啪啪无遮挡十八禁网站| 青春草亚洲视频在线观看| 国产在线一区二区三区精| 视频区欧美日本亚洲| 日韩制服骚丝袜av| 精品卡一卡二卡四卡免费| 嫩草影视91久久| 欧美日韩中文字幕国产精品一区二区三区 | 国产免费一区二区三区四区乱码| 少妇粗大呻吟视频| 亚洲成av片中文字幕在线观看| 精品熟女少妇八av免费久了| 国产成人免费无遮挡视频| 午夜两性在线视频| 在线精品无人区一区二区三| 国产黄频视频在线观看| 91精品伊人久久大香线蕉| 天天躁夜夜躁狠狠躁躁| 久久久久精品人妻al黑| 亚洲人成电影免费在线| 色婷婷久久久亚洲欧美| 久久国产精品男人的天堂亚洲| 午夜福利在线观看吧| 色视频在线一区二区三区| 免费在线观看影片大全网站| 国产一区二区三区综合在线观看| 亚洲五月婷婷丁香| 搡老乐熟女国产| 久久久久国产精品人妻一区二区| 十八禁网站网址无遮挡| 精品国产乱码久久久久久小说| 精品卡一卡二卡四卡免费| 久久亚洲精品不卡| 精品熟女少妇八av免费久了| 亚洲欧美一区二区三区久久| 午夜影院在线不卡| 一级片免费观看大全| 高潮久久久久久久久久久不卡| 中文字幕制服av| 久久av网站| 亚洲人成电影观看| 成人18禁高潮啪啪吃奶动态图| 一边摸一边做爽爽视频免费| 美女高潮喷水抽搐中文字幕| 日韩大片免费观看网站| 老司机亚洲免费影院| 国产精品香港三级国产av潘金莲| 在线观看免费日韩欧美大片| 亚洲欧洲精品一区二区精品久久久| 色老头精品视频在线观看| 国产精品国产av在线观看| 久久精品久久久久久噜噜老黄| 欧美 亚洲 国产 日韩一| 日韩电影二区| 超色免费av| 啦啦啦在线免费观看视频4| 亚洲第一av免费看| 久久九九热精品免费| 国产成人免费无遮挡视频| 少妇被粗大的猛进出69影院| 一级片免费观看大全| 久久午夜综合久久蜜桃| 亚洲伊人色综图| 电影成人av| 亚洲欧美一区二区三区久久| 国产主播在线观看一区二区| 欧美日韩一级在线毛片| a级毛片黄视频| 在线观看免费视频网站a站| a在线观看视频网站| svipshipincom国产片| 精品少妇一区二区三区视频日本电影| 中文字幕人妻丝袜一区二区| 91麻豆av在线| 91字幕亚洲| 欧美精品一区二区大全| 亚洲av电影在线观看一区二区三区| 亚洲av成人不卡在线观看播放网 | 欧美精品av麻豆av| 精品人妻1区二区| 这个男人来自地球电影免费观看| 一级毛片女人18水好多| 午夜影院在线不卡| 欧美日韩视频精品一区| 国产成人a∨麻豆精品| 午夜福利视频在线观看免费| 男人操女人黄网站| 欧美激情久久久久久爽电影 | 黄色视频在线播放观看不卡| 久久久久精品人妻al黑| 操美女的视频在线观看| 久久人人爽人人片av| 国产精品.久久久| 2018国产大陆天天弄谢| 91字幕亚洲| 欧美在线一区亚洲| 国精品久久久久久国模美| 午夜免费观看性视频| 日本wwww免费看| 秋霞在线观看毛片| 国产成人精品在线电影| 国产精品欧美亚洲77777| 飞空精品影院首页| 日本a在线网址| 淫妇啪啪啪对白视频 | 少妇裸体淫交视频免费看高清 | 大香蕉久久成人网| 搡老熟女国产l中国老女人| 久久久欧美国产精品| 麻豆av在线久日| 中文字幕高清在线视频| 在线永久观看黄色视频| 日本一区二区免费在线视频| 国产精品久久久久久精品电影小说| 国产一区二区激情短视频 | 精品国产一区二区三区四区第35| 亚洲精品久久久久久婷婷小说| 91九色精品人成在线观看| 国产三级黄色录像| 黄片播放在线免费| 亚洲精品久久成人aⅴ小说| 制服诱惑二区| 在线十欧美十亚洲十日本专区| 精品少妇黑人巨大在线播放| 女警被强在线播放| 亚洲欧美精品综合一区二区三区| 亚洲av日韩精品久久久久久密| 午夜成年电影在线免费观看| 99国产极品粉嫩在线观看| 国产熟女午夜一区二区三区| 侵犯人妻中文字幕一二三四区| www.av在线官网国产| 久久ye,这里只有精品| 亚洲成人免费av在线播放| 制服诱惑二区| 国产精品免费视频内射| 国产成人精品在线电影| 91大片在线观看| tube8黄色片| 亚洲人成电影免费在线| 老司机在亚洲福利影院| 亚洲av日韩精品久久久久久密| 777米奇影视久久| 午夜视频精品福利| 天堂中文最新版在线下载| 啦啦啦在线免费观看视频4| 国产主播在线观看一区二区| 日韩一卡2卡3卡4卡2021年| 国产无遮挡羞羞视频在线观看| 免费看十八禁软件| 手机成人av网站| 狠狠狠狠99中文字幕| 亚洲国产欧美网| 美女高潮喷水抽搐中文字幕| 精品久久久久久电影网| 国产精品99久久99久久久不卡| 午夜福利在线免费观看网站| 久久ye,这里只有精品| 免费少妇av软件| 国产成人av教育| 国产成人一区二区三区免费视频网站| 黄网站色视频无遮挡免费观看| 国产免费福利视频在线观看| 久久99热这里只频精品6学生| 国产欧美日韩一区二区三 | 精品第一国产精品| 日韩制服丝袜自拍偷拍| 大香蕉久久成人网| 欧美精品亚洲一区二区| 国产精品自产拍在线观看55亚洲 | 国产三级黄色录像| 久久99热这里只频精品6学生| 男女无遮挡免费网站观看| 日本撒尿小便嘘嘘汇集6| 激情视频va一区二区三区| 永久免费av网站大全| 一区二区三区乱码不卡18| av有码第一页| 淫妇啪啪啪对白视频 | 青草久久国产| 侵犯人妻中文字幕一二三四区| 大型av网站在线播放| 精品少妇一区二区三区视频日本电影| 最黄视频免费看| 极品少妇高潮喷水抽搐| √禁漫天堂资源中文www| 欧美激情 高清一区二区三区| 啦啦啦在线免费观看视频4| 99精国产麻豆久久婷婷| 国产精品一区二区精品视频观看| 国产欧美日韩一区二区精品| 日本91视频免费播放| 法律面前人人平等表现在哪些方面 | 在线av久久热| 国产精品国产av在线观看| 免费在线观看日本一区| 欧美日韩亚洲综合一区二区三区_| 国产精品久久久久久精品电影小说| 免费人妻精品一区二区三区视频| 日韩制服丝袜自拍偷拍| 熟女少妇亚洲综合色aaa.| 亚洲,欧美精品.| 精品高清国产在线一区| 国产成人精品无人区| 一区二区三区激情视频| 菩萨蛮人人尽说江南好唐韦庄| 国产真人三级小视频在线观看| av天堂久久9| av有码第一页| 亚洲一区中文字幕在线| 欧美日韩视频精品一区| av在线老鸭窝| 丰满迷人的少妇在线观看| 国产精品九九99| 欧美 亚洲 国产 日韩一| 又大又爽又粗| 久久精品国产a三级三级三级| av网站在线播放免费| 人妻一区二区av| 男男h啪啪无遮挡| 大型av网站在线播放| 热re99久久精品国产66热6| 久久这里只有精品19| 日韩精品免费视频一区二区三区| 亚洲精品久久午夜乱码| 精品免费久久久久久久清纯 | 欧美日韩av久久| 欧美日韩视频精品一区| 国产精品麻豆人妻色哟哟久久| 欧美精品av麻豆av| 亚洲av电影在线进入| 曰老女人黄片| 中文字幕人妻熟女乱码| 男人添女人高潮全过程视频| 亚洲欧美日韩高清在线视频 | 欧美一级毛片孕妇| 中文字幕人妻丝袜一区二区| 国产成+人综合+亚洲专区| 久久久久国内视频| 午夜免费成人在线视频| av一本久久久久| 亚洲av日韩精品久久久久久密| 亚洲一区中文字幕在线| 99精品久久久久人妻精品| 国产日韩欧美在线精品| 午夜免费鲁丝| 亚洲欧美精品自产自拍| 精品国产超薄肉色丝袜足j| 色婷婷av一区二区三区视频| 蜜桃在线观看..| 亚洲国产欧美网| 午夜福利在线观看吧| 欧美+亚洲+日韩+国产| 法律面前人人平等表现在哪些方面 | 欧美国产精品一级二级三级| 国产三级黄色录像| 精品福利观看| 久久精品久久久久久噜噜老黄| 男女无遮挡免费网站观看| 脱女人内裤的视频| cao死你这个sao货| 18禁观看日本| 欧美少妇被猛烈插入视频| 欧美日韩中文字幕国产精品一区二区三区 | 亚洲成人手机| 久久天躁狠狠躁夜夜2o2o| 午夜视频精品福利| 国产男女内射视频| 国产精品.久久久| 国产精品一区二区精品视频观看| 热99re8久久精品国产| 两个人免费观看高清视频| 男男h啪啪无遮挡| 一级,二级,三级黄色视频| 亚洲av成人不卡在线观看播放网 | 视频区欧美日本亚洲| 97人妻天天添夜夜摸| 久久精品久久久久久噜噜老黄| h视频一区二区三区| 午夜福利视频精品| av超薄肉色丝袜交足视频| 丝瓜视频免费看黄片| 久久久精品94久久精品| 成在线人永久免费视频| 丰满饥渴人妻一区二区三| 亚洲免费av在线视频| 欧美激情 高清一区二区三区| 亚洲精品国产av成人精品| 青草久久国产| 国产xxxxx性猛交| 日日摸夜夜添夜夜添小说| av超薄肉色丝袜交足视频| 国产精品影院久久| 久久影院123| 黄频高清免费视频| 欧美日韩亚洲高清精品| 国产亚洲午夜精品一区二区久久| 午夜福利免费观看在线| 亚洲三区欧美一区| 在线av久久热| 欧美日本中文国产一区发布| 手机成人av网站| 亚洲国产av影院在线观看| 另类亚洲欧美激情| 一本大道久久a久久精品| 男人操女人黄网站| 亚洲精品久久成人aⅴ小说| 99久久综合免费| 一区二区三区精品91| 天天影视国产精品| 在线观看免费高清a一片| 又大又爽又粗| 国产一区有黄有色的免费视频| 亚洲 欧美一区二区三区| 亚洲精华国产精华精| 精品亚洲乱码少妇综合久久| 成人18禁高潮啪啪吃奶动态图| 精品免费久久久久久久清纯 | 激情视频va一区二区三区| 婷婷丁香在线五月| 亚洲av日韩精品久久久久久密| 麻豆乱淫一区二区| 成年女人毛片免费观看观看9 | 丝瓜视频免费看黄片| 国产日韩一区二区三区精品不卡| 亚洲七黄色美女视频| 好男人电影高清在线观看| 中国国产av一级| 亚洲av日韩在线播放| 韩国精品一区二区三区| 黑人欧美特级aaaaaa片| 国产福利在线免费观看视频| 日韩制服丝袜自拍偷拍| 亚洲情色 制服丝袜| 男女免费视频国产| 国产深夜福利视频在线观看| 久久人人爽av亚洲精品天堂| 最近最新中文字幕大全免费视频| 欧美+亚洲+日韩+国产| 欧美久久黑人一区二区| 99国产精品一区二区蜜桃av | 悠悠久久av| 国产高清videossex| 国产亚洲精品久久久久5区| 悠悠久久av| 成人影院久久| 久久99一区二区三区| 国产在线一区二区三区精| 人人妻人人添人人爽欧美一区卜| 国产精品 国内视频| 天天躁狠狠躁夜夜躁狠狠躁|