• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Two-Order Approximate and Large Stepsize Numerical Direction Based on the Quadratic Hypothesis and Fitting Method

    2020-05-21 05:45:32XiaoliYinChunmingLiandYuanZhang
    IEEE/CAA Journal of Automatica Sinica 2020年3期

    Xiaoli Yin, Chunming Li, and Yuan Zhang

    Abstract—Many effective optimization algorithms require partial derivatives of objective functions, while some optimization problems’ objective functions have no derivatives. According to former research studies, some search directions are obtained using the quadratic hypothesis of objective functions. Based on derivatives, quadratic function assumptions, and directional derivatives, the computational formulas of numerical first-order partial derivatives, second-order partial derivatives, and numerical second-order mixed partial derivatives were constructed. Based on the coordinate transformation relation, a set of orthogonal vectors in the fixed coordinate system was established according to the optimization direction. A numerical algorithm was proposed, taking the second order approximation direction as an example. A large stepsize numerical algorithm based on coordinate transformation was proposed. Several algorithms were validated by an unconstrained optimization of the two-dimensional Rosenbrock objective function. The numerical second order approximation direction with the numerical mixed partial derivatives showed good results. Its calculated amount is 0.2843% of that of without second-order mixed partial derivative. In the process of rotating the local coordinate system 360°, because the objective function is more complex than the quadratic function, if the numerical direction derivative is used instead of the analytic partial derivative, the optimization direction varies with a range of 103.05°. Because theoretical error is in the numerical negative gradient direction, the calculation with the coordinate transformation is 94.71% less than the calculation without coordinate transformation. If there is no theoretical error in the numerical negative gradient direction or in the large-stepsize numerical optimization algorithm based on the coordinate transformation, the sawtooth phenomenon occurs. When each numerical mixed partial derivative takes more than one point, the optimization results cannot be improved. The numerical direction based on the quadratic hypothesis only requires the objective function to be obtained, but does not require derivability and does not take into account truncation error and rounding error. Thus, the application scopes of many optimization methods are extended.

    I. Introduction

    OPTIMIZATION methods are widely used in various research fields and play an important role in scientific and technological problems [1]–[7]. At present, the research of optimization methods is mainly divided into multi-objective[8]–[10], intelligence, genetic algorithm [10], [11], clustering[12] methods, and so on. Classical multidimensional optimization methods are also studied [13].

    Many efficient optimization methods need derivative and partial derivative information of the objective function, such as the linear fitting derivative method, high-order second approximate fixed-point method, tangent intersection point method, negative gradient direction method, zigzag line negative gradient direction method (blind- walking method),second approximate direction method [14], linear fitting gradient method [15], various conjugate direction methods[16]–[18], the optimal available direction method (feasible direction method), half step method [19], and so on. These methods are not useful in optimization problems where the objective function is not differentiable. Numerical differentiation algorithms [20] may allow these methods to work.

    Comparing the above algorithms, the second approximation direction is the best, and the direction of the negative gradient is most commonly used. The negative gradient direction is the fastest local descent direction, and usually deviates from the extreme value point in the whole design space. The conjugate directions constructed on the basis of the negative gradient direction have better optimization efficiency [21], [22].

    The numerical differential algorithm can not only determine the numerical partial derivative, but also obtain the numerical directional derivative in the general direction.

    Looking at the teaching contents of the course “mechanical optimum design” or “optimization methods” and the research literature of optimization methods, and at classical optimization methods, it can be determined that they are all proposed under the quadratic objective function hypothesis and then applied to the general objective function. Although there is an optimization direction based on finite differences[23], the truncation error and rounding error are emphasized while the quadratic hypothesis is neglected due to the fitting effect of the objective function. By representing each derivative in the system of differential equations by its finite difference approximation [24], the quadratic hypothesis is ignored. The research based on the quadratic hypothesis and fitting method is more general.

    Classical optimization methods are not proposed to deal with large-scale matrices. However, some large-scale sparse matrix problems borrow optimization ideas, such as the graph coloring problem [25] and nonlinear equations solving method[26]. Although document [25] deals with the optimization direction based on finite differences, it is far from the research in this paper.

    According to the characteristics of the optimization method,the numerical partial derivative and the numerical directional derivative of the objective function are constructed by using the numerical differentiation algorithm under the quadratic function assumption. On this basis, the optimization effect of the numerical second approximate direction and the local coordinate negative gradient direction is studied.

    II. Construction Method of Numerical First and Two Order Partial Derivatives Based on the Quadratic Hypothesis

    The optimization methods with a clear direction of optimization are put forward under the unimodal assumption.The optimization method based on the quadratic function is universal. There is no universal significance in the study of higher order objective functions. In the design space, the objective function is assumed to be quadratic in any direction,so there are three undetermined coefficients. Take the current pointx(k)as the center and taking pointx(k?1)and pointx(k+1)by a microstep ?sat two sides. Assume that the objective function values of the three points aref(k?1),f(k)andf(k+1),respectively. ?sshall be reasonably determined according to the computer’s expressive ability and termination condition value. Three equations can be obtained from the above three points. Thus the above three parameters are obtained so that the quadratic function of the hypothesis is known. The partial derivatives at pointsx(k?1),x(k)andx(k+1)are shown as follows:

    The above three formulas are consistent with the results obtained by the Lagrange interpolation function method. The second partial derivatives at the three points are equal.

    Equation (4) is consistent with the Newton center difference formula. The numerical gradient and the numerical second partial derivative at the pointx(k)can be obtained from the above 4 formulas, but the second mixed partial derivative cannot be obtained.

    III. Construction Method of Numerical Directional Derivative

    The directional derivative is the rate of change of a function along a certain direction, which is a generalization of the concept of a partial derivative. Letsbe the direction fromx(k)tox(k+1), wherescan be expressed as follows:

    The gradient function of the objective function at pointx(k)is as follows:

    The directional derivative can be regarded as the point product of the gradient vector and the unit directional vector[27], equal to the length of the vector ?f(x) projected ons.

    where cos(?f(x),s) is the cosine of the angle betweenands.kis the optimal serial number. The minute step alongsis decomposed into the minute step alongNcoordinate axes. Then, according to the definition of the partial derivative (7) can be obtained.

    According to the definition of directional derivative change rate, the numerical directional derivative atx(m), the midpoint ofx(k+1)andx(k)is as follows:

    Ifx(k+1)is sufficiently close tox(k), the numerical directional derivative at pointx(k)can be expressed approximately by (8). Calculate the objective function valuef(m)of pointx(m); then the numerical directional derivatives at pointx(k)are as follows:

    A linear equation of gradient vectors can be obtained from(7) and (8). By searching forNpoints near the current pointx(0)and establishing the above equations, a set ofN-ary linear equations can be obtained, and the numerical gradient of the objective function can be solved.

    The numerical negative gradient direction obtained from(10) has some theoretical errors due to the approximate directional derivative of (8). According to the quadratic function hypothesis, the linear equations can be obtained from(7) and (9).

    According to the method of taking points above, the coefficient matrix (10) and (11) is a real diagonal matrix, is non-singular, and its rank isN. It is almost impossible to find the same objective function at each point as that at pointx(0),so (10) and (11) have a unique nonzero solution. In the same way, if the first partial derivative of the objective function is known,Npoints can be found near the current pointx(0), and the aboveN-ary linear equations can be obtained, so that the second partial derivative matrix of the objective function can be solved by the least square method.

    IV. Construction Method of the Numerical Directional Derivative

    Fig. 1. The point drawing for obtaining mixed partial derivative by numerical differentiation.

    Then the second mixed partial derivative is obtained from the following equation.

    According to the quadratic function hypothesis, the relationship between the two coordinates has six undetermined coefficients, and the five points in Fig. 1 can not determine these coefficients, so a numerical point must be added. If we take pointfon the bisector of two coordinate axes, then according to (2) at the midpointgofcf, the partial derivative ofxiis as follows:

    The numerical second mixed partial derivative is derived from the assumption that the objective function is quadratic,so the results of (13) and (14) should be the same.

    The parameters (3), (4), (13) and (14) are not set according to the specific objective function, but determined, accurate and unchanged. Without the value of theoretical analysis,there is no need to discuss truncation error and rounding error.

    V. The Construction Method of a Local Orthogonal Coordinate System for Optimization

    If a local coordinate system is established in the design space, the partial derivative of the objective function can hardly be obtained, but the numerical partial derivative can be obtained using the numerical differential algorithm as in the global coordinate system. After optimizing in a certain direction, it is better to search in the direction of orthogonality again. An orthogonal direction vector group is set up whose last direction is the optimization direction. The orthogonal vector group can be regarded as the coordinate vector group of the local coordinate system. The coordinate vector group is then the coordinate transformation matrixAbetween the two coordinate systems. If the origin of the local coordinate system isx(0)in the fixed coordinate system, then a pointx′in the local coordinate system is shown in the global coordinate system as follows:

    The vectors′in the local coordinate system is shown in the global coordinate system as shown below:

    Set the previous optimization directionSas the last vector αN. Let the largest element ofS(the element with the largest absolute value) be the elementi. Then in theN×Nunit matrix, it corresponds to the column vectori. In order, the column vector of the unit matrix acts as other vectors; thus we can obtain the following linear independent vectors

    The modified Schmitt orthogonalization method is used to perform orthogonalization [28], [29]. We use the following definitions and change vector to unitization vectors with two norm

    Among them,kis an undetermined coefficient. According to the orthogonality between vectors, the following formula can be obtained. We then change the vectors to unitization vectors with two norm

    The square matrix composed of unitized vectors β1,β2,...,βNis the coordinate transformation matrix of (15)and (16).

    The orthogonal vector group can also makeSthe last coordinate axis of the local coordinate system after coordinate transformation, so as to obtain the last column elements ofAaccording to (16). Then, according to the type of coordinate transformation, other elements ofAcan be obtained.

    For a non-quadratic objective function, the optimization direction obtained from the numerical gradient and the second partial derivative matrix constructed by 1, 2, 3 in different local coordinate systems is different.

    VI. Large Stepsize Numerical Optimization Algorithm Based on Coordinate Transformation

    The second approximate direction directly points to the extreme point of the quadratic objective function. Such a numerical direction is good. Another new algorithm is as follows:

    In order to make full use of the optimization effect obtained by the detection point, after each detection of the best point,the best point is taken as the current pointx, and the originalxis taken as the reserve pointx(i)(iis the coordinate axis serial number). If neither probe point is good enough, halve the step size and continue to probe until the step size is less than the termination condition value. The next best point is taken as the reserved pointx(i), and then, from the current point (the best point) detect along the next coordinate axis.After sequential detection alongNcoordinate axes, the numerical negative gradient or numerical second approximation direction is constructed. A step optimization is achieved after one dimensional optimization until the current pointxis close enough to the reserve pointx(i).

    Due to the large distance between the detection points, the numerical first partial derivatives and second partial derivatives have large errors. Instead of calculating the mixed partial derivative and one-dimensional optimization in the new direction, the backward half-step searching algorithm[30] based on the blind path-finding idea is used to find better points according to the current step size. The optimization scheme is shown in Fig. 2.

    Because of the continuity of the feasible region, the initial point can be chosen arbitrarily. If the step size is too large, the optimization direction may deviate too much from the negative gradient direction and the optimization effect would not be good; if the step size is too small, it causes a sawtooth phenomenon because it is too close to the negative gradient direction. Therefore, the initial step size should be determined according to the specific optimization problem, and a better value can be determined by trial calculation.

    VII. Example Verification

    A. The Two-Dimensional Example

    The two-dimensional Rosenbrock function is the four square function, and it is one of the most difficult examples to test the optimization algorithm with. The unconstrained optimization problem whose objective function is as follows:

    The gradient of the objective function is as follows:

    The second partial derivative matrix of the objective function is as follows:

    The objective function consists of two components, each of which has a minimum value of zero. A system of two variables and quadratic equations can be obtained, and the extreme points and extremes can be obtained as follows:

    Fig. 2. The program diagram of large step numerical algorithm based on coordinate transformation.

    Take the initial point and the micro step length as follows:

    The objective function value, numerical gradient and second partial derivative matrix at this point are as follows:

    The relative errors between the numerical solutions obtained from (2), (4), (13), and (14) and the analytical solutions are all less than 1 0?5.

    In this paper, the numerical second approximate direction algorithm and the large step numerical algorithm based on coordinate transformation are realized by theClanguage. The blind searching method [31] is used in one dimensional optimization.

    B. The Test of Numerical Second Approximate Direction Method

    The optimal direction is obtained by constructing the numerical diagonal second partial derivative matrix from (4).After one-dimensional optimization is performed 3955 times,the objective function is calculated 156 517 times. The optimal point and optimal solution are obtained as follows:

    The optimization process is shown in Fig. 3, solid line 1.

    Fig. 3. The seeking process of the numerical second-order approximate direction method.

    The optimal direction is obtained by constructing the numerical symmetric second partial derivative matrix from(4), (13), (14). After one-dimensional optimization is performed 11 times, the objective function is calculated 445 times, and the optimal point and solution are obtained as follows:

    With the numerical second mixed partial derivative, the optimization calculation amount is 0.2843% of without it. The optimization process is shown in the gem-shaped point and dashed line 2 in Fig. 3. The result is similar to that of the analytical second approximate direction method.

    A local coordinate system is established for each optimization.After one-dimensional optimization is performed 244 times, the objective function is called 9884 times, and the optimal point and solutions are obtained as follows:

    Its optimization process is shown in Fig. 3, dotted line 3.Coordinate transformation is not suitable.

    C. The Test of the Constructing Search Direction in Local Coordinate System

    In order to test the influence of coordinate transformation on the numerical optimization direction, a local coordinate system is established withx(0)as the origin, and the local coordinate system is rotated counterclockwise 360°.

    The numerical direction derivative of the fixed coordinate system is used as the numerical partial derivative of the local coordinate system to construct the optimization direction. In the above rotation process, the trajectory of the numerical optimization direction in the local coordinate system is shown in the plus point and dashed line 1 in Fig. 4. It is not evenly distributed, indicating that it is not a fixed direction. In the fixed coordinate system, the trajectory is not a fixed direction,but varies within the range 103.05°, as shown by the asterisk point and dotted line 2 in Fig. 4. It is clear that the calculated results are consistent with those of the 5 analyses.

    Fig. 4. Numerical direction trajectories in the rotation process of local coordinate system.

    D. Test of the Large Step Numerical Optimization Algorithm

    The initial step size, the minimum initial step length, and the termination condition value are as follows:

    In the first optimization round, the following points are detected along theNaxis:

    According to (27), the analytic gradient of the objective function at this point is as follows:

    The numerical gradient obtained by (9) is as follows:

    The difference between them is quite large. Then we obtain theN(2) midpoint, and the numerical gradient obtained by(11) is as follows:

    Using (9), after 1437 detection rounds, the objective function is called 89 155 times, and the optimal point and solutions are obtained as follows:

    The optimization process is shown in Fig. 5, dotted line 1.Adopting (11), the objective function must be called 87 065 times in 1726 detection rounds.

    Fig. 5. The search process of the large step numerical algorithm based on coordinate transformation.

    Then the numerical optimization direction of (4) and (21) is as follows:

    After 1430 detection rounds, we call the objective function 70 621 times to get the optimal point and solution.

    The optimization process is shown in Fig. 5, solid line 2.

    The numerical gradient is obtained by means of 5 coordinate transformation and (9). After 144 detection rounds,we call the objective function 4717 times to get the optimal point and solution

    The optimization process is shown in the dashed line 3 of Fig. 5. Compared with the black dotted line, the computation amount is reduced by 94.71%. Due to the large step size, the numerical negative gradient direction obtained by (9) has a large error. If the numerical direction is rotated 90, after 243 rounds of detection, the objective function is called 7872 times, and the optimal point and solution are obtained as follows:

    The calculation amount is not large, which is consistent with the analysis of 7.3. In the process of approaching, the extremum point is skipped over for several times, but not arrived at step by step. After the numerical negative gradient direction is constructed by (11) and compared with the dotted line 1, the optimal solution is 2.4764×10?15after 14 975 detection rounds and 944 941 calls to the objective function.There is a very serious zigzag phenomenon because it is too close to the analytic negative gradient direction.

    E. The Three-Dimensional Example

    The three-dimensional Rosenbrock function is one of the most difficult multi-dimensional examples to test the optimization algorithm with. The unconstrained optimization problem whose objective function is as follows:

    The gradient of the objective function is as follows:

    The second partial derivative matrix of the objective function is as follows:

    The objective function consists of 4 components, each of which has a minimum value of 0. A system of three variable quadratic equations can be obtained, and the extreme points and extremes can be obtained as follows:we take the initial point as follows:

    Adopting the analytic 2 order approximation direction, after 16 rounds of one-dimensional optimization, the objective function is calculated 614 times, and the optimal point and solution are obtained as follows:

    Adopting the numerical two order approximation direction,if we do not calculate the numerical mixed partial derivative,after 446 rounds of one-dimensional optimization, the objective function is calculated 20 038 times, and the optimal point and solution are obtained as follows:

    Using (13) or (14) to calculate the numerical mixed partial derivative, after 16 rounds of one-dimensional optimization,the objective function is calculated 614 times, and the optimal point and solution are obtained as follows:

    Taking more than one point for each mixed partial derivative, it is then computed by using (13) or (14). After 16 rounds of one-dimensional optimization, the objective function is calculated 660 times, and the optimal point and solution are obtained as follows:

    VIII. Discussion

    The mathematical deduction in this paper is concise and easy to understand. A lot of tedious deduction processes are omitted. The key derivation steps and principles are explained in the text and it is similar to many classical numerical differentiation formulas (difference formula).

    The purpose of this paper is to broaden the scope of existing optimization methods. It should test the approximation between numerical and analytical solutions. If the two approaches are close enough, the optimization effect of the original optimization method will not be affected. The Rosenbrock function is four square, with a curved canyon. At the bottom of the canyon, the objective function contour is close to the parallel, and there are no characteristics of the quadratic function at all. The optimization methods with the optimal direction are all based on the quadratic objective function hypothesis. Some optimization algorithms cannot obtain the effective direction at the bottom of the valley.Therefore, this example is one of the most difficult examples to test the optimization algorithm with. The successful verification of the numerical example is sufficient to prove the optimization effect of numerical optimization direction.

    Conjugate directions needNto expand enough in optimization space, and more than three-dimensional examples are needed to verify the optimization effect. The research in this paper does not involve the verification of conjugate directions. Two-dimensional examples can not only illustrate the problem, but also visually represent the optimization process. Even if the numerical algorithm of the conjugate direction is studied, a three-dimensional numerical example is not necessary, because if the analytical direction is not good, the numerical direction will not be good.

    For multimodal objective functions, different initial points may have different optimization results. For a single peak objective function, the optimization algorithm with a certain direction will not yield different optimal points because of the different initial points. The initial point of the example in this paper is taken on a flat slope on the side of the canyon, which is far from the extreme value point. It not only checks the speed of searching to the canyon bottom, but also checks the effect of approaching the extreme value point at the canyon bottom.

    In this paper, we compare different algorithms using computation and optimization results, and we show that it is not necessary to use an elegant intelligent algorithm. This algorithm is not comparable with an evolutionary algorithm or swarm intelligence optimization algorithm. The latter is an optimization method based on other theories, and does not require gradient information such as in the Powell method[32]. However, there is no concept of optimizing direction; it is an algorithm that seeks order in disorder; it is an algorithm that explores necessity in occasionality. The algorithm has obvious advantages for multi-modal feasible regions, but for single-modal feasible regions, it is not as effective as an algorithm with a definite direction. The numerical direction constructed by the numerical method is not more accurate than that obtained by the analytical method, and both are suitable for a single-peak feasible region.

    IX. Conclusion

    Through the verification of several numerical algorithms with two-dimensional Rosenbrock objective function optimization, the following conclusions can be drawn:

    1) The numerical formulas for calculating the first, second and second mixed partial derivatives have higher accuracy.The numerical first partial derivatives are in agreement with the results obtained by the method of constructing Lagrange interpolation functions. Thus, these numerical formulas are correct. Because it is difficult to obtain second partial derivatives, research in optimization methods that only need the gradient of objective function have been explored. This paper continues the research in this area. At present, there are some research directions involving using numerical direction to improve the accuracy of optimization. The research based on the quadratic hypothesis also ends this direction.

    2) The numerical second approximate direction method with numerical second mixed partial derivatives is the best.

    3) The optimization direction obtained in a local coordinate system has strong randomness, and the large step numerical algorithm requires it.

    4) Based on the quadratic hypothesis and fitting method, the numerical algorithm proposed in this paper allows the derivatives, gradients, second partial derivatives and second mixed partial derivatives of the objective function to remain unresolved, which makes optimization methods that require this information more applicable. There is no derivable objective function for optimization problems in many engineering fields. This study provides more optimization methods for solving these optimization problems.

    5) Because constrained optimization problems are often encountered in control processes, future research on constrained optimization methods may become the focus of our research.

    51午夜福利影视在线观看| 欧美高清性xxxxhd video| 女生性感内裤真人,穿戴方法视频| 日韩欧美精品v在线| 国产亚洲精品久久久久久毛片| 内射极品少妇av片p| 精品国内亚洲2022精品成人| 精品久久久久久成人av| 精品一区二区三区视频在线| 最近最新免费中文字幕在线| 最近中文字幕高清免费大全6 | 国产精品一区二区三区四区免费观看 | 亚洲美女视频黄频| 国产精品免费一区二区三区在线| 国产成年人精品一区二区| 亚洲18禁久久av| 欧美一区二区国产精品久久精品| 女人十人毛片免费观看3o分钟| 午夜免费激情av| 欧美黄色淫秽网站| 久久久精品大字幕| 中文资源天堂在线| 757午夜福利合集在线观看| 久久久久久久久久黄片| 午夜两性在线视频| 人妻夜夜爽99麻豆av| 亚洲精品日韩av片在线观看| 国产高清激情床上av| 国产伦精品一区二区三区视频9| 久久精品国产自在天天线| 久久国产乱子伦精品免费另类| 午夜免费成人在线视频| 久久久久亚洲av毛片大全| 18禁黄网站禁片免费观看直播| 久久国产乱子伦精品免费另类| 国产高清激情床上av| 丰满人妻一区二区三区视频av| 亚洲美女视频黄频| 丰满人妻一区二区三区视频av| 五月玫瑰六月丁香| 欧美精品国产亚洲| 波多野结衣高清作品| 日本精品一区二区三区蜜桃| 久久久久久久久大av| 亚洲av.av天堂| 午夜a级毛片| 亚洲自偷自拍三级| 国内揄拍国产精品人妻在线| 3wmmmm亚洲av在线观看| 精品人妻1区二区| 国产一级毛片七仙女欲春2| 国产一区二区三区视频了| 最好的美女福利视频网| 日本五十路高清| 日本五十路高清| 久久精品夜夜夜夜夜久久蜜豆| 国产一区二区三区视频了| 身体一侧抽搐| 大型黄色视频在线免费观看| 国产三级在线视频| 91午夜精品亚洲一区二区三区 | 亚洲不卡免费看| 国产精华一区二区三区| 亚洲欧美日韩卡通动漫| 能在线免费观看的黄片| 国产69精品久久久久777片| 国产又黄又爽又无遮挡在线| 久久久久免费精品人妻一区二区| 国内精品一区二区在线观看| 色综合婷婷激情| 特级一级黄色大片| 成人特级黄色片久久久久久久| 日本免费一区二区三区高清不卡| aaaaa片日本免费| 免费看美女性在线毛片视频| 国产精品一区二区三区四区久久| 一a级毛片在线观看| 变态另类丝袜制服| 亚洲av.av天堂| 高清日韩中文字幕在线| 综合色av麻豆| 一区二区三区免费毛片| 亚洲人成电影免费在线| 久久久久精品国产欧美久久久| 国产欧美日韩精品一区二区| 网址你懂的国产日韩在线| 久久国产乱子免费精品| 变态另类丝袜制服| 一夜夜www| 高清毛片免费观看视频网站| 欧美色视频一区免费| x7x7x7水蜜桃| 真实男女啪啪啪动态图| h日本视频在线播放| 国产精品亚洲一级av第二区| 日韩精品青青久久久久久| 亚洲av免费在线观看| a级毛片a级免费在线| 欧美激情在线99| 欧美色视频一区免费| 人妻久久中文字幕网| 成人美女网站在线观看视频| 激情在线观看视频在线高清| 亚洲成人中文字幕在线播放| 欧美日韩乱码在线| 一本一本综合久久| 亚洲男人的天堂狠狠| 成人永久免费在线观看视频| 国产一级毛片七仙女欲春2| 我的老师免费观看完整版| 欧美bdsm另类| 久久久久久久久大av| 久久精品综合一区二区三区| 99热6这里只有精品| 欧美日韩国产亚洲二区| 欧美午夜高清在线| 亚洲人成伊人成综合网2020| 成年版毛片免费区| 亚州av有码| 亚洲人与动物交配视频| 国产白丝娇喘喷水9色精品| 国产蜜桃级精品一区二区三区| 国产成人影院久久av| 国产 一区 欧美 日韩| 99久久久亚洲精品蜜臀av| or卡值多少钱| 级片在线观看| 99国产极品粉嫩在线观看| 日韩欧美免费精品| 亚洲天堂国产精品一区在线| av在线观看视频网站免费| 狠狠狠狠99中文字幕| 色尼玛亚洲综合影院| 久久香蕉精品热| 天堂√8在线中文| 色播亚洲综合网| 桃色一区二区三区在线观看| 久久草成人影院| 国产精品不卡视频一区二区 | 亚洲熟妇熟女久久| 桃红色精品国产亚洲av| 欧美黄色淫秽网站| 国产熟女xx| 小说图片视频综合网站| 午夜福利在线观看免费完整高清在 | 国产成人aa在线观看| 如何舔出高潮| 欧美绝顶高潮抽搐喷水| h日本视频在线播放| 全区人妻精品视频| 成人国产一区最新在线观看| 深夜a级毛片| 久久久久久久久大av| 长腿黑丝高跟| 最近中文字幕高清免费大全6 | 国产精品久久久久久久电影| 亚洲欧美日韩无卡精品| 亚洲精品成人久久久久久| 宅男免费午夜| 亚洲av日韩精品久久久久久密| 日韩av在线大香蕉| 中文字幕精品亚洲无线码一区| 亚洲avbb在线观看| 99热这里只有是精品50| 国产av麻豆久久久久久久| 人人妻人人看人人澡| 久久久久久久久久成人| 人人妻,人人澡人人爽秒播| 久久人人爽人人爽人人片va | 亚洲av免费高清在线观看| 成人鲁丝片一二三区免费| 久久久久久久久中文| 国产爱豆传媒在线观看| 亚洲国产精品sss在线观看| 日本精品一区二区三区蜜桃| 久久久久久久精品吃奶| 别揉我奶头~嗯~啊~动态视频| 丰满人妻熟妇乱又伦精品不卡| netflix在线观看网站| 小说图片视频综合网站| 自拍偷自拍亚洲精品老妇| av天堂在线播放| 亚洲精品粉嫩美女一区| 丰满乱子伦码专区| 男人和女人高潮做爰伦理| x7x7x7水蜜桃| av黄色大香蕉| 99热这里只有是精品50| 国产精品乱码一区二三区的特点| 又粗又爽又猛毛片免费看| 黄片小视频在线播放| 亚洲七黄色美女视频| 好男人在线观看高清免费视频| 99久久精品国产亚洲精品| 老女人水多毛片| 久久久久免费精品人妻一区二区| 91久久精品电影网| 久久久久久久久大av| 午夜福利免费观看在线| 亚洲精品久久国产高清桃花| 99热这里只有是精品在线观看 | 熟女电影av网| av中文乱码字幕在线| 国产私拍福利视频在线观看| 国产精品自产拍在线观看55亚洲| 国产人妻一区二区三区在| 午夜久久久久精精品| 赤兔流量卡办理| av福利片在线观看| 亚洲精品亚洲一区二区| 婷婷色综合大香蕉| 亚洲18禁久久av| 两个人视频免费观看高清| 中国美女看黄片| 能在线免费观看的黄片| 极品教师在线视频| 窝窝影院91人妻| 亚洲一区二区三区不卡视频| 婷婷精品国产亚洲av| 免费看a级黄色片| 草草在线视频免费看| 亚洲最大成人av| 99riav亚洲国产免费| 18禁黄网站禁片午夜丰满| 国产aⅴ精品一区二区三区波| 欧美绝顶高潮抽搐喷水| 在线观看av片永久免费下载| 久久人人精品亚洲av| 一个人免费在线观看电影| 亚洲最大成人手机在线| av黄色大香蕉| 欧美日韩瑟瑟在线播放| 精品人妻1区二区| 麻豆国产97在线/欧美| 性色avwww在线观看| 蜜桃亚洲精品一区二区三区| 91在线观看av| 最好的美女福利视频网| 国产精品三级大全| 中国美女看黄片| 国产综合懂色| 成人特级黄色片久久久久久久| 久久人人精品亚洲av| 丰满乱子伦码专区| 日韩亚洲欧美综合| 国产69精品久久久久777片| 国产欧美日韩精品一区二区| 精品福利观看| 日本与韩国留学比较| 国产亚洲欧美在线一区二区| 欧美黑人巨大hd| 性欧美人与动物交配| 波多野结衣巨乳人妻| 久久香蕉精品热| 精品人妻视频免费看| 超碰av人人做人人爽久久| 欧美成人免费av一区二区三区| 99久久九九国产精品国产免费| 欧美bdsm另类| 午夜福利在线观看吧| 欧美国产日韩亚洲一区| 亚洲 欧美 日韩 在线 免费| 亚洲精品色激情综合| 亚洲性夜色夜夜综合| 首页视频小说图片口味搜索| 国内精品久久久久精免费| 不卡一级毛片| 久久久久久大精品| 性欧美人与动物交配| 国产伦在线观看视频一区| 2021天堂中文幕一二区在线观| 99在线人妻在线中文字幕| 桃红色精品国产亚洲av| 级片在线观看| 熟妇人妻久久中文字幕3abv| 久久精品人妻少妇| 欧美精品国产亚洲| www.www免费av| 亚洲不卡免费看| 亚洲va日本ⅴa欧美va伊人久久| 在线观看av片永久免费下载| 五月玫瑰六月丁香| 一本久久中文字幕| 国产综合懂色| 在线十欧美十亚洲十日本专区| 国产男靠女视频免费网站| 在线观看美女被高潮喷水网站 | 亚洲成人久久爱视频| 国产成人啪精品午夜网站| 亚洲最大成人中文| 亚洲国产精品sss在线观看| 制服丝袜大香蕉在线| 亚洲精品456在线播放app | 成人精品一区二区免费| 国产亚洲精品久久久久久毛片| 国产私拍福利视频在线观看| 亚洲精品影视一区二区三区av| 日本撒尿小便嘘嘘汇集6| 赤兔流量卡办理| 亚洲人成网站在线播放欧美日韩| 草草在线视频免费看| 日韩欧美三级三区| 日日干狠狠操夜夜爽| 99久国产av精品| 91九色精品人成在线观看| 国产精品久久久久久亚洲av鲁大| 极品教师在线免费播放| 国产成人aa在线观看| 啦啦啦韩国在线观看视频| av在线蜜桃| 天堂√8在线中文| 亚洲人成伊人成综合网2020| 99国产精品一区二区三区| 色在线成人网| 我要搜黄色片| 一级av片app| 一本综合久久免费| 精品人妻1区二区| 国产精品爽爽va在线观看网站| 好男人在线观看高清免费视频| 国产大屁股一区二区在线视频| 精品福利观看| 亚洲精品成人久久久久久| 国内毛片毛片毛片毛片毛片| 在线免费观看不下载黄p国产 | 一级av片app| 婷婷精品国产亚洲av| 97超视频在线观看视频| 成人av在线播放网站| 窝窝影院91人妻| 波多野结衣高清无吗| 免费观看人在逋| 亚洲人成网站在线播| 999久久久精品免费观看国产| 国产欧美日韩精品亚洲av| 成人性生交大片免费视频hd| 小蜜桃在线观看免费完整版高清| 97人妻精品一区二区三区麻豆| 男女之事视频高清在线观看| 一级av片app| 亚洲经典国产精华液单 | 成人亚洲精品av一区二区| 女同久久另类99精品国产91| 九色成人免费人妻av| 简卡轻食公司| 国产熟女xx| 久久精品国产自在天天线| 中亚洲国语对白在线视频| 久久伊人香网站| .国产精品久久| 国产高清视频在线观看网站| 午夜精品在线福利| av在线老鸭窝| 91字幕亚洲| 色综合站精品国产| 两人在一起打扑克的视频| 日日摸夜夜添夜夜添av毛片 | 小蜜桃在线观看免费完整版高清| 大型黄色视频在线免费观看| 久久久精品欧美日韩精品| 国产v大片淫在线免费观看| 婷婷亚洲欧美| 日韩欧美三级三区| 麻豆成人午夜福利视频| 亚洲中文日韩欧美视频| 午夜福利18| 毛片一级片免费看久久久久 | 一本久久中文字幕| 欧美极品一区二区三区四区| 精品福利观看| 黄片小视频在线播放| 一本久久中文字幕| 淫妇啪啪啪对白视频| 97碰自拍视频| 欧美性感艳星| 亚洲一区高清亚洲精品| 伊人久久精品亚洲午夜| www日本黄色视频网| 欧美成人性av电影在线观看| 十八禁国产超污无遮挡网站| 国产一区二区三区视频了| 少妇裸体淫交视频免费看高清| 中文在线观看免费www的网站| 久久欧美精品欧美久久欧美| 亚洲aⅴ乱码一区二区在线播放| 国产综合懂色| 成人国产一区最新在线观看| 久久久久久久精品吃奶| 精品久久久久久成人av| 国产乱人伦免费视频| 精品久久国产蜜桃| 亚洲美女搞黄在线观看 | 岛国在线免费视频观看| 国产精品永久免费网站| 亚洲成人免费电影在线观看| 综合色av麻豆| 男人狂女人下面高潮的视频| 999久久久精品免费观看国产| or卡值多少钱| 成年女人毛片免费观看观看9| 精品不卡国产一区二区三区| 国产精品免费一区二区三区在线| 国产成人aa在线观看| 免费看光身美女| 丰满人妻一区二区三区视频av| 日韩有码中文字幕| 在线十欧美十亚洲十日本专区| 久久精品国产清高在天天线| 日韩有码中文字幕| 免费看光身美女| 美女高潮喷水抽搐中文字幕| 99热6这里只有精品| 国产精品98久久久久久宅男小说| 在线免费观看的www视频| 久久99热6这里只有精品| 波野结衣二区三区在线| 舔av片在线| 久久精品国产99精品国产亚洲性色| 国产熟女xx| 色噜噜av男人的天堂激情| 97热精品久久久久久| 免费在线观看日本一区| 午夜免费成人在线视频| 性插视频无遮挡在线免费观看| av专区在线播放| 久久精品久久久久久噜噜老黄 | 欧美+日韩+精品| 亚洲美女黄片视频| 亚洲性夜色夜夜综合| av专区在线播放| 老熟妇仑乱视频hdxx| 男女视频在线观看网站免费| 久久久久久久亚洲中文字幕 | bbb黄色大片| 网址你懂的国产日韩在线| 亚洲最大成人av| 老师上课跳d突然被开到最大视频 久久午夜综合久久蜜桃 | 麻豆国产97在线/欧美| 精品人妻一区二区三区麻豆 | 国产成人欧美在线观看| 欧美三级亚洲精品| 午夜福利在线观看吧| 中文资源天堂在线| 亚洲自偷自拍三级| 日韩欧美国产在线观看| eeuss影院久久| 97超级碰碰碰精品色视频在线观看| av在线蜜桃| 18美女黄网站色大片免费观看| 少妇熟女aⅴ在线视频| 欧美性感艳星| 99久久精品一区二区三区| 一a级毛片在线观看| 国产成人av教育| 精品福利观看| 18禁在线播放成人免费| 成人美女网站在线观看视频| www日本黄色视频网| 国产成人aa在线观看| 一区二区三区激情视频| 亚洲狠狠婷婷综合久久图片| 国产伦一二天堂av在线观看| 村上凉子中文字幕在线| 日本在线视频免费播放| 有码 亚洲区| 国产成+人综合+亚洲专区| 老鸭窝网址在线观看| 亚洲国产高清在线一区二区三| 国产精品美女特级片免费视频播放器| 日韩成人在线观看一区二区三区| 又爽又黄a免费视频| 国产成人福利小说| 成人永久免费在线观看视频| 成人无遮挡网站| 1000部很黄的大片| 男人狂女人下面高潮的视频| 宅男免费午夜| 午夜日韩欧美国产| 欧美黑人巨大hd| 免费在线观看影片大全网站| 久久久国产成人精品二区| 99久久精品热视频| 看片在线看免费视频| 日本a在线网址| 国产老妇女一区| 一进一出抽搐动态| 男人舔女人下体高潮全视频| 精品久久久久久久久久免费视频| 18禁在线播放成人免费| 欧美日韩综合久久久久久 | 国产精品美女特级片免费视频播放器| 五月伊人婷婷丁香| av在线老鸭窝| 国内毛片毛片毛片毛片毛片| 午夜两性在线视频| 99久久精品国产亚洲精品| 草草在线视频免费看| 精品国内亚洲2022精品成人| а√天堂www在线а√下载| 夜夜看夜夜爽夜夜摸| x7x7x7水蜜桃| 丰满的人妻完整版| 国产精品人妻久久久久久| 免费搜索国产男女视频| 我的老师免费观看完整版| 成人鲁丝片一二三区免费| 亚洲av.av天堂| 在线国产一区二区在线| 免费搜索国产男女视频| 日韩免费av在线播放| 91麻豆av在线| 天堂√8在线中文| 两性午夜刺激爽爽歪歪视频在线观看| 国产亚洲精品久久久com| 国产三级在线视频| 久久人妻av系列| 听说在线观看完整版免费高清| av福利片在线观看| 中文字幕精品亚洲无线码一区| 热99在线观看视频| a级一级毛片免费在线观看| 欧美激情在线99| 每晚都被弄得嗷嗷叫到高潮| 成熟少妇高潮喷水视频| 九色成人免费人妻av| 757午夜福利合集在线观看| 真人一进一出gif抽搐免费| 精品久久久久久,| 亚洲精品乱码久久久v下载方式| 草草在线视频免费看| 日韩欧美国产在线观看| 日韩欧美三级三区| 国产精品一区二区免费欧美| 国产高清视频在线观看网站| 中文字幕久久专区| 国产av不卡久久| 99精品在免费线老司机午夜| 亚洲精品色激情综合| 夜夜躁狠狠躁天天躁| 亚洲18禁久久av| 午夜精品一区二区三区免费看| 国产高清视频在线观看网站| 国产成人a区在线观看| 男女床上黄色一级片免费看| 老女人水多毛片| 亚洲综合色惰| 日本三级黄在线观看| av国产免费在线观看| 亚洲av不卡在线观看| 男女做爰动态图高潮gif福利片| 中文在线观看免费www的网站| 久久中文看片网| 亚洲天堂国产精品一区在线| 亚洲自偷自拍三级| 成人欧美大片| 听说在线观看完整版免费高清| 日韩国内少妇激情av| 99久久无色码亚洲精品果冻| 亚洲经典国产精华液单 | 赤兔流量卡办理| 成年女人永久免费观看视频| 可以在线观看毛片的网站| 91麻豆精品激情在线观看国产| 国产久久久一区二区三区| 日韩欧美精品v在线| x7x7x7水蜜桃| 日韩精品青青久久久久久| 中文字幕精品亚洲无线码一区| 久久热精品热| 亚洲成人久久爱视频| 久久热精品热| 51国产日韩欧美| 精品人妻偷拍中文字幕| 热99re8久久精品国产| 在线观看av片永久免费下载| 能在线免费观看的黄片| 成人av一区二区三区在线看| 久9热在线精品视频| 悠悠久久av| 能在线免费观看的黄片| 亚洲成人久久性| 久久久久性生活片| 日韩成人在线观看一区二区三区| 亚洲人成网站在线播| 丰满的人妻完整版| 99热精品在线国产| 久久久精品欧美日韩精品| 757午夜福利合集在线观看| 亚洲五月天丁香| 国产极品精品免费视频能看的| 午夜精品久久久久久毛片777| 神马国产精品三级电影在线观看| 成人av在线播放网站| 在线观看av片永久免费下载| 能在线免费观看的黄片| 国产精品久久久久久亚洲av鲁大| 在线免费观看不下载黄p国产 | h日本视频在线播放| 桃红色精品国产亚洲av| 夜夜看夜夜爽夜夜摸| 久久九九热精品免费| 每晚都被弄得嗷嗷叫到高潮| 久9热在线精品视频| 久久热精品热| 老司机午夜十八禁免费视频| 精品人妻1区二区| 特大巨黑吊av在线直播| 日韩中字成人| 精品99又大又爽又粗少妇毛片 | 十八禁国产超污无遮挡网站| 少妇的逼好多水| 日韩欧美在线乱码| 在线观看美女被高潮喷水网站 | 一a级毛片在线观看| 99久久成人亚洲精品观看|