• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Efficient Differential Evolution for Truss Sizing Optimization Using AdaBoost Classifier

    2023-01-22 09:00:32TranHieuNguyenandAnhTuanVu

    Tran-Hieu Nguyenand Anh-Tuan Vu

    Hanoi University of Civil Engineering,Hanoi,100000,Vietnam

    ABSTRACT Design constraints verification is the most computationally expensive task in evolutionary structural optimization due to a large number of structural analyses that must be conducted.Building a surrogate model to approximate the behavior of structures instead of the exact structural analyses is a possible solution to tackle this problem.However,most existing surrogate models have been designed based on regression techniques.This paper proposes a novel method,called CaDE,which adopts a machine learning classification technique for enhancing the performance of the Differential Evolution (DE) optimization.The proposed method is separated into two stages.During the first optimization stage, the original DE is implemented as usual, but all individuals produced in this phase are stored as inputs of the training data.Based on design constraints verification, these individuals are labeled as“safe” or “unsafe” and their labels are saved as outputs of the training data.When collecting enough data, an AdaBoost model is trained to evaluate the safety state of structures.This model is then used in the second stage to preliminarily assess new individuals,and unpromising ones are rejected without checking design constraints.This method reduces unnecessary structural analyses,thereby shortens the optimization process.Five benchmark truss sizing optimization problems are solved using the proposed method to demonstrate its effectiveness.The obtained results show that the CaDE finds good optimal designs with less structural analyses in comparison with the original DE and four other DE variants.The reduction rate of five examples ranges from 18 to over 50%.Moreover, the proposed method is applied to a real-size transmission tower design problem to exhibit its applicability in practice.

    KEYWORDS Structural optimization;machine learning;surrogate model;differential evolution;AdaBoost classifier

    1 Introduction

    Nowadays,global competition has posed a big challenge for industrial companies to offer quality products at lower prices.The situation is similar for the field of construction where the contractors are frequently pressured to reduce costs while maintaining high standards.In this case, optimizing the weight of structures is seen as an effective strategy to gain a competitive advantage.Traditionally,the structure is often designed manually following the“trial-and-error”method.The final structural solution received from this way is deeply influenced by the designers’experience and intuition.Therefore, in order to help the designers find the best solution, the optimization algorithms were applied to the design process and have been constantly developed.

    Generally, optimization algorithms are categorized into two groups:gradient-based group and metaheuristic group.The gradient-based optimization finds the minimum by going in the direction opposite sign of the derivative of the function.Hence,the speed of convergence is seriously influenced by the starting point,and the gradient-based optimization may fall to the local minimum.Additionally,the function is sometimes difficult or impossible to calculate the derivative.Contrary to gradient-based methods,metaheuristics do not require the calculation of derivative value.Instead,these algorithms use the randomization principle to find the optimum.By searching for a large set of solutions spreading over the design space,metaheuristics can easily escape a local optimum and eventually reach a global optimum.Because of their advantages like easy implementation as well as robustness,metaheuristics are increasingly applied to civil engineering problems.Some outstanding algorithms that are widely used in structural optimization can be listed here such as Genetic Algorithm (GA) [1,2], Evolution Strategies(ES)[3,4],Differential Evolution(DE)[5,6],Particle Swarm Optimization(PSO)[7,8].Some recently proposed algorithms have shown competitive performance,for example,Artificial Bee Colony(ABC) [9], Harmony Search (HS) [10,11], Teaching-Learning-Based Optimization (TLBO) [12,13],Colliding Bodies Optimization(CBO)[14,15].

    Despite many advantages,metaheuristics have one common drawback of requiring a large number of fitness evaluations.It should be highlighted that for structural optimization problems,the objective function can be quickly determined but the computation of constraints is very costly due to involving the structural analysis.Conducting a high number of structural analyses is the most time-consuming task when optimizing structures.Therefore, reducing the number of structural analyses becomes an important goal.To overcome this challenge,many efforts have been made to improve the performance of existing algorithms[16–19].Besides,using surrogate models instead of computationally expensive structural analyses is a potential solution[20–28].In recent decades,due to the fast growth of Artificial Intelligence(AI),especially Machine Learning(ML),the topic regarding ML-based surrogate-assisted metaheuristic algorithms has attracted interest from researchers.

    Based on the literature, there are several strategies for managing surrogate models.The most straightforward way is to build surrogate models based on the data obtained from the original fitness evaluations and the whole models will be used during the optimization process.Since 1997,Papadrakakis et al.[20] employed feed-forward neural networks (NNs) to predict the response of structures.These NNs were combined with the ES algorithm for optimizing shapes and sections of structures.To cope with the error of NNs,a correction technique was proposed,in which the allowable constraint values were adjusted proportionally to the maximum error for the testing dataset.The research was then extended to large-scale 3D roof trusses [21].Similarly, structural analyses were replaced by NNs during the GA-based optimization process[22].In more detail,three kinds of NN including radial basis function (RBF) network, generalized regression (GR) network, and counterpropagation(CP)network were used for optimizing 25-bar truss structure as well as 1300-bar dome structure.The obtained results showed that the RBF networks outperformed the other two networks.In 2019,Taheri et al.[23]combined the RBF neural networks and the ABC algorithm for solving the size,shape,and topology optimization problems of steel lattice towers.Next,the adaptive fuzzy neural inference system(ANFIS)network is used instead of structural analyses carried out by MSTOWER software when optimizing transmission towers[24].Nguyen et al.[25]utilized two deep NNs separately to estimate the stresses and displacements when optimizing truss structures.Recently,Mai et al.[26]also used deep NNs to reduce the computation cost of the design optimization of geometrically nonlinear trusses.This strategy has been proven to be time-effective but the designs were not the true optimal designs due to the inaccuracy of surrogate models.

    To avoid falling a false optimum, both the exact functions and the approximate functions are used simultaneously.For example,in[27],the combination of NN and GA was adapted to optimize transmission towers based on this strategy.The optimization process employed in this study comprises two stages:in the first stage, the RBF networks were used as structural analyzers until the stopping criterion and in the second stage, the exact structural analyses were conducted.Chen et al.[28]proposed a hybrid approach,in which the near-optimal point was located by the ES algorithm with the help of NN-based surrogate models,and the sequential quadratic programming algorithm was then used to find the exact optimal point.

    These existing studies have been used a similar approach,in which the surrogate models have been constructed by regression techniques with the aim of predicting the degree of constraint violation of candidates.While this approach is very intuitive and has been successfully applied in a lot of studies,this is not the only one.An alternative approach is to use the classification technique to build surrogate models.In the literature, the classification-based surrogate models have been successfully applied to solve optimization problems.For instance,Kramer et al.[29]used a classification model for predicting the feasibility of the solutions.However,a feasible solution can be classified as an infeasible solution due to the poor quality model,leading to the rejection of potential solutions.Therefore,the constraint boundary is shifted to the infeasible region by a small distance to limit this error.Another study[30]proposed a probability which is called the influence coefficient to prevent the misclassification of the surrogate model.In 2018,Wang et al.[31]introduced a technique in which the boundary is adjusted to decrease the possibility of classifying feasible solutions to be infeasible.

    Although classification techniques have been used previously,the proposed method in the present work is different.Instead of using an additional parameter as in former works, this study employs a classification surrogate model in conjunction with the direct comparison of the objective function values.In detail,a classifier will be built to identify whether the structural solution satisfies the design constraints or not.When optimizing the weight of trusses,if a new solution is predicted to violate the design constraints and its weight is heavier than that of the parent solution,it will be discarded without conducting structural analysis.In this way,some unnecessary structural analyses will be reduced and consequently,the computational time will be shortened.

    The classification technique used in this study is Adaptive Boosting (as known as AdaBoost)developed by Freund et al.[32].The AdaBoost classifier is combined with the DE algorithm to form an efficient optimizer called CaDE.AdaBoost is selected because it has been proven as a powerful algorithm for the task of classifying the structural safety of trusses [33].In addition, the choice of the DE algorithm is motivated by its simplicity and robustness.The proposed method CaDE is first examined for the weight optimization of truss structures.

    The remainder of this paper is arranged as follows.A brief introduction of the DE algorithm and the AdaBoost classifier are given in Section 2.The truss sizing optimization is presented in Section 3.1.Next,the detail of the proposed method CaDE is described in Section 3.2.Computational experiments are conducted in Section 4,in which,six benchmark trusses are optimized according to the proposed method to show its efficiency.In addition, a real-size transmission tower is also designed using the proposed method to exhibit its applicability in practice.Finally,Section 5 concludes this paper.

    2 Background Information

    2.1 Differential Evolution Algorithm

    Differential Evolution is a metaheuristic algorithm that was firstly introduced by Storn et al.[34]in 1995.Thanks to its simplicity,this algorithm is widely applied to many fields such as communication,transportation, aerospace, civil engineering, etc.Since the first introduction, DE has been taken a top rank in many competitions relating to evolutionary computation [35,36].Particularly, some studies that compare the performance of metaheuristic algorithms in structural optimization have been conducted[37].The results show that DE outperforms the remaining algorithms.

    DE is designed based on the natural evolutionary mechanism.According to the Darwinian theory,the main driving forces of evolution are mutation and natural selection.Similarly,DE consists of four basic steps as shown in Fig.1,in which the initialization step is conducted one time at the first of the optimization process while three remaining steps are repeated till a stopping criterion is satisfied.

    Figure 1:Main steps of differential evolution

    The optimization process begins with an initialization of a population pop0containingNPindividuals.Each individual representing a candidate of the optimization problem is aD-dimensional vector:

    In order to find the global optimum,the initial population should be spread over the search space as much as possible.The most widely used method for generating the initial population is to use the uniform distribution.In detail, thejth component of theith vector can be initialized based on this formula:

    where:xjminandxjmaxare the lower and upper bounds for thejth component of theith vector;rand[0,1]is a uniformly distributed random number in the range[0,1].

    Secondly, DE createsNPmutant vectors vi(t)corresponding toNPtarget vector xi(t)via the mutation operator.The classical mutation operator is to add a scaled difference vector between two random vectors to the third vector.It can be written in the mathematical form as follows:

    where:r1 ≠r2 ≠r3 ≠iare three integers that are randomly chosen from the range[1,NP];tis the current generation number;Fis the scaling factor.

    The above mutation strategy is called “DE/rand/1”where the term “rand”means random and the number“1”denotes that only one difference vector is added.Besides“DE/rand/1”,four mutation strategies that were proposed by Storn et al.[34]are listed below:

    in which:xbest(t)is the best individual of the current population.

    In the crossover step,a trial vector ui(t)is constructed by taking some components from xi(t)and the rest from vi(t).It can be expressed as:

    where:ui,j(t),vi,j(t)andxi,j(t)are thejth component of ui(t),vi(t),and xi(t),respectively;Kis a random integer in[1,D]to ensure that ui(t)differs xi(t)by getting at least one component from vi(t);andCris called the crossover rate.

    Based on natural selection, only the fittest survive.For minimization problems, the fitness is measured through the value of the objective function.Only the one having the lower value of the objective function will be kept for the next generation.It can be described as follows:

    where:f(.)is the objective function that needs to be optimized.

    2.2 AdaBoost Classifier

    There exist many powerful machine learning classification models such as Support Vector Machine or Neural Networks.The abovementioned algorithms are categorized into the single learning group.However,there exists another group of algorithms,called ensemble methods,in which several classifiers are combined with the aim of improving the accuracy.Ensemble methods have been proven to be very effective for classification tasks.In this study, a boosting method, called AdaBoost, is employed.The idea of this algorithm is to train iteratively several weak learners using the weighted training data, and then incorporate them into the final classifier [32].The weak learners can be any single learning algorithm,however,the most commonly used one is the Decision Tree.In the following section,the background theory of the AdaBoost classifier is presented.

    Generally,AdaBoost produces several classifiers,called weak classifiers,each of them has a proper weight based on its accuracy.The final prediction is the combination of the output of weak classifiers using the weight majority voting technique.The more accurate classifier will contribute more impact on the final prediction by assigning a higher weight.In addition,AdaBoost keeps a set of weights over the training data.Initially, the weights of all samples are set equally.After an iteration, the weights of samples that were misclassified by the previous weak classifier are increased.It ensures that those samples will receive the attention of the weak classifier on the next iteration.The AdaBoost algorithm is illustrated in Fig.2.

    Figure 2:Illustration of the AdaBoost algorithm

    Suppose the training dataset isDB={(x1,y1),(x2,y2),...,(xn,yn)}in which xi,i=1,2,...,nis ap-dimensional input vector containingpfeatures;yi∈{+1,-1}is the output;andnis the number of samples in the training dataset.

    AdaBoost begins by initializing sample weights uniformly as:

    where:w1,iis the initial weight of the sample(xi,yi).

    Then a process of four basic steps is iteratively conducted forTtimes.On iterationt, the weak classifier isht,and the weight vector is wt={wt,1,wt,2,...,wt,n}.

    Step 1:the classifierhtis fit to the weighted training data:

    where:DecisionTree(.)indicates a single tree classifier.

    Step 2:the error of the classifier on this iteration is calculated as follows:

    where:1(ht(xi) ≠yi) is the function that returns 1 for correctly classified samples while 0 for misclassified samples.

    Step 3:The weight of the weak classifierhtcan be determined based on its accuracy as follows:

    Step 4:the weights of samples are updated according to the following rule:

    AfterTsuch iterations,Tweak classifiersht,t=1, 2,...,Tare obtained.The final output is the weighted sum of the outputs of these weak classifiers:

    where:sign(x)is a function that returns-1 ifx <0,+1 ifx >0,0 ifx=0.

    3 The Proposed Method CaDE

    3.1 Truss Sizing Optimization Using DE

    Optimizing a truss structure is a specific constrained optimization problem where the objective function is frequently the weight of the truss structure and the constraints relate to the structural behavior like stress, displacement, or vibration conditions.Three categories of truss optimization problems include size, shape, and topology optimization, in which size optimization is widely used in practical design.For size optimization,the member cross-sectional areas are considered as design variables.The mathematical formulation of this type of optimization problem is expressed as follows:

    where:A is a vector containingngdesign variables;W(A)is the objective function which is the weight of the whole truss structure in this case;ρis the density;Aidenotes a design variable which is the cross-sectional area of members belong to theith group;Ljis the length of thejth member;ngis the number of groups;nm(i)is the number of members belong to theith group;gk(A)is thekth constraint;ncis the number of constraints;Ai,minandAi,maxare the lower and upper bounds of the design variableAi,respectively.

    To apply metaheuristic algorithms which are originally designed for unconstrained optimization,the most widely-used technique is to use the fitness function instead of the objective function:

    in which:Fit(A) is the fitness function; andP(A) is the penalty function which is added whenever any constraint is violated;cv(A)is the degree of constraint violation that is defined as the sum of all constraint violations:

    3.2 Accelerating DE with AdaBoost Classifier

    The objective function of structural optimization problems,for example,the weight of structures,can be easily calculated by multiplying the volume of all members with their material density.However,the determination of the fitness function is very costly due to conducting a computationally expensive structural analysis.As result,the computation time of metaheuristic optimization tasks that require thousands or even millions of fitness function evaluations becomes prohibitively high.

    Surrogate models have been frequently constructed to simulate the response of structures, and they have been used to quickly evaluate the fitness of individuals instead of exact structural analyses.Although the accuracy of surrogate modeling techniques has been greatly improved in recent years,there is always an error between the predictions and the ground truth due to the complexity of the structural response as well as the limitation of training data.Therefore,the direct replacement of exact structural analysis with surrogate models may lead to a false optimum[20–25].

    The proposed method employs surrogate models in a different manner.The idea behind the proposed method is based on an observation that a trial vector that violates any constraint and has a larger objective function value is largely worse than its target vector and it cannot survive the selection.Detecting and eliminating these worse trial vectors during the optimizing process can save many useless structural analyses.The objective function of structural optimization problems can be quickly determined but the constraint verification is very costly due to performing a structural analysis.It leads to the idea of constructing a surrogate model that can quickly evaluate the safety state of the structure.Using regression models to estimate the degree of constraint violation for this task is not necessary.Instead,a classification model that can distinguish whether or not the individual satisfies constraints is sufficient.

    The concept of the proposed method is the application of a machine learning classifier to identify trial vectors that are most likely worse than these target vectors during the optimization process.In this work,the AdaBoost algorithm is employed for that purpose and the proposed method is called CaDE.The pseudo-code of the CaDE is described in Algorithm 1.

    Algorithm 1.Pseudo-code of the CaDE 1: Set g=0,DB=null,NP,F,Cr,n_iter1,max_iter#Phase 1:Building model 2: Initialize a population pop0={xi(0)|i=1,2, ...,NP}using LHS technique 3: for each xi(0)in pop0 do 4: Evaluate xi(0)5: Assign yi(0)according to Eq.(19)(Continued)

    Note:where: g is the current iteration number; DB is the database for training model; NP is the size of the population; F is the scaling factor;Cr is the crossover rate;n_iter1 is the numbers of iterations for the model building phase;max_iter is the number of iterations of the whole optimization process;Mutation,Crossover,and Selection are three basic operators of the DE as described in Section 2.1;C(.)is the AdaBoost classifier;W(.)is the objective function.

    Generally,the CaDE has two phases:the model building phase and the model employing phase.The CaDE algorithm begins by initializing a population ofNPvectors pop0= {xi(0)|i=1, 2,...,NP}.It is noted that the initial population is generated using the Latin Hypercube Sampling(LHS)technique because of its good space-filling capability.Each vector xi(0)is checked design constraints,and the labelyi(0)is assigned using the following formula:

    As a result,the initial datasetDBconsists ofNPdata samples{(xi(0),yi(0))|i=1,2,...,NP}.Next,three operators(Mutation,Crossover,and Selection)are carried out sequentially as the original DE algorithm.All trial vectors ui(g)produced by the Mutation and Crossover operators are checked design constraints and assigned a labelyi(g)according to Eq.(19).All obtained results(ui(g),yi(g))are also saved into the databaseDB.The loop is terminated after(n_iter1-1)iterations.The final dataset comprisesn_iter1×NPdata samples.At the end of this phase, an AdaBoost classificationCmodel has been trained by the datasetDBas presented in Section 2.2.

    In the model employing phase,each trial vector ui(g)is preliminarily evaluated using the classification modelC.At this time,two conditions are applied in determining which trial vectors to go to the selection step or not.In more detail,a trial vector is rejected without exact evaluation when it has two conditions simultaneously:(i)it is predicted to violate design constraints(ypred_i(g)=-1),and(ii)its objective function value is larger than that of the corresponding target vector(W(ui(g))>W(wǎng)(xi(g))).Otherwise,the constraint violation of the trial vector is evaluated using the structural analysis and the Selection operator is then conducted as usual.The model employing phase lasts until the pre-defined number of iterationmax_iterreaches.

    4 Numerical Experiments

    Numerical experiments are performed on a personal computer with CPU Intel Core i5 2.5 GHz,8 Gb RAM,and Windows 10 64 bit OS.The DE algorithm and the CaDE algorithm are developed by the authors using the Python programming language,meanwhile,the AdaBoost classifier is built using the open-source library scikit-learn[38].The Python in-house codes for finite element analysis of truss structures are developed based on the direct stiffness method described in[39].

    4.1 Illustrative Example

    A three-bar truss example is carried out to illustrate the application of the CaDE algorithm to structural optimization.Although this structure is simple and the constraints can be directly evaluated,the stresses of the members are still determined through the finite element analyses in this experiment.This problem is used as an illustrative example because it has only two variables and it can be easily visualized.

    The geometry, loading, and boundary conditions of the structure are presented in Fig.3.The objective of this example is to find the cross-sectional areasA1andA2to minimize the weight of the truss while satisfying the condition that the stresses in all bars do not exceed the allowable stress.This problem is stated as follows:

    in which:the applied forceP=2 kN;L=100 cm;σkis the stress in thekth member;and the allowable stress[σ]=2 kN/cm2.

    Figure 3:Three-bar truss structure

    As mentioned above,the objective functionW(A)should be transformed into the fitness function.In this example,the fitness function is defined asFit(A)=W(A)(1+cv(A))and it is plotted in Fig.4.

    Figure 4:Visualization of the fitness function of the three-bar truss structure

    The DE algorithm starts by initializing an initial population pop0ofNPindividuals.In this example, the number of individuals of a population is set toNP=20.The initial individuals are represented together with the 2-D contour plot of the fitness function as shown in Fig.5a.Then,three operators including mutation,crossover,and selection are sequentially performed to produce a population pop1for the next iteration with two optimization parameters as follows:F= 0.8,Cr=0.9.The manner in which the mutant vector vi(0)and the trial vector ui(0)are produced from the target vector xi(0)is shown in Fig.5b.If the fitness function value of the trial vector ui(0)is smaller than that of the target vector xi(0),the trial vector ui(0)is selected to be the target vector xi(1)of the next population.Otherwise,the trial vector ui(0)is discarded and the target vector xi(0)is kept for the next iteration.

    Figure 5:Visualization of three first operators of the DE algorithm

    The above process is repeated until the stopping condition is reached.In this example,the stopping condition is set to be after 100 iterations.The populations at several iterations are shown in Fig.6.It can be seen that all individuals of the population tend to move closer to the global optimal position during the optimization process.At the final iteration,almost all individuals are concentrated at the global optimal position.The optimal design found by the DE algorithm is as follows:A1=0.789 cm2andA2=0.408 cm2.The number of structural analyses that must be conducted isnSA=2000 times.

    The optimization parametersF,Cr,max_iterof the CaDE are kept the same as in the DE.The number of iterations for the first phase is set ton_iter1=5, meaning the number of samples in the datasetDBis equal ton_iter1×NP=5×20=100.Fig.7 presents all samples collected in the first phase,where vectors that are labeled+1 are represented as white dots while vectors that are labeled-1 are represented as red dots.Based on the obtained dataset,an AdaBoost classifier is built as the green dashed curve shown in Fig.7.

    In the second phase,the decision of performing exact structural analyses depends on the prediction of the AdaBoost classifier.Fig.8 illustrates the optimization process of the proposed method at the 5thiteration, in which all white dots represent the target vectors xi(5)of the current population After conducting two operators including mutation and crossover, the machine learning classifier is employed to predict the label of the trial vector ui(5).At this time,one of three possible cases can occur as follows:

    Figure 6:Improvement of the population during the DE optimization process

    Figure 7:Building an AdaBoost classifier based on historical constraint evaluations

    Case 1:If the predicted labelypred_i(5)=+1,the trial individual ui(5)is then checked design constraints using exact structural analysis.These individuals are represented as blue dots in Fig.8;

    Case 2:If the predicted labelypred_i(5)=-1 and the objective function of the trial vectorW(ui(5))is smaller than or equal to that of the target vectorW(xi(5)), it is also checked design constraints using exact structural analysis.These individuals are represented as yellow dots in Fig.8;

    Case 3:If the predicted labelypred_i(5)=-1 andW(ui(5))>W(wǎng)(xi(5)),it is eliminated without performing the exact structural analysis.These individuals are marked as magenta dots in Fig.8.

    Figure 8:Illustration of the implementation of the proposed method CaDE

    After reaching the stopping condition,the optimal cross-sectional areas of truss members found by the CaDE areA1=0.789 cm2andA2=0.408 cm2which are similar to the results found by the original DE.However,by applying the AdaBoost classifier,the number of exact structural analyses is greatly reduced.The optimization is independently performed 30 times and the statistical results show that the proposed method CaDE requires an average of 1770 structural analyses while the original DE algorithm needs 2000 structural analyses.

    4.2 Five Benchmark Truss Problems

    4.2.1 Description of the Problems

    Five well-known truss sizing optimization problems collected from the literature including the 10-bar planar truss,the 17-bar planar truss,the 25-bar spatial truss,the 72-bar spatial truss,and the 200-bar planar truss are carried out in this section.

    The first example considers a simple truss containing 10 members as illustrated in Fig.9.This structure is made of the material having the modulus of elasticityE=10,000 ksi and the densityρ=0.1 lb/in3.This problem has 10 design variables A={Ai,i=1,2,...,10}where 0.1 in2≤Ai≤33.5 in2is the cross-sectional areas ofith member.Two vertical loads acting at node(2)and node(4)have the same magnitudeP=100 kips.Stresses inside members must be lower than 25 ksi for both compression and tension.Displacements of nodes(1),(2),(3),(4)are limited within±2 in.

    Figure 9:Layout of the 10-bar truss

    The second example considers a 17-bar planar truss as presented in Fig.10.This structure is fabricated from a material that has the modulus of elasticityE=30,000 ksi and the material densityρ=0.268 lb/in3.Each member in this structure has its own cross-sectional areas and there are 17 design variables in total A = {0.1 in2≤Ai≤33.5 in2,i=1, 2,..., 17}.A vertical force of 100 kips acts at node(9)of the truss.The allowable stresses in both tension and compression areσallow=50 ksi and the allowable displacements along the X and Y directions areδallow=2 in.

    Figure 10:Layout of the 17-bar truss

    In the third example,a spatial truss containing 25 members as presented in Fig.11 is considered.This structure is made of the same material as the 10-bar truss (E=10,000 ksi andρ=0.1 lb/in3).Members of this structure are classified into 8 groups where members belonging to the same group are assigned the same cross-sectional area that varies between 0.01 in2to 3.5 in2.This structure is optimized according to both stress constraint and displacement constraint.The allowable stresses are presented in Table 1 and the permissible displacement of all nodes is±0.35 in.The loading data are shown in Table 2.

    Figure 11:Layout of the 25-bar truss

    Table 1:Member groups and the corresponding allowable stress values for the 25-bar truss

    Table 2:Loading data for the 25-bar truss(kips)

    The layout of the 72-bar spatial truss investigated in the fourth example is displayed in Fig.12.Members are classified into 16 groups as described in Table 3.The minimum and maximum areas of all members are 0.01 in2and 4.0 in2.Two load cases acting on this structure are as follows:(LC1)two horizontal forces of 5 kips along the X-axis and Y-axis and a vertical force of 5 kips along the opposite direction of the Z-axis act at node(17);(LC2)four vertical forces of 5 kips along the opposite direction of the Z-axis act at four nodes(17),(18),(19),(20),respectively.The design constraints of this problem include stress constraint and displacement constraint.The allowable stresses of all members are±25 ksi while the allowable displacements along the X and Y directions are±0.25 in.

    Figure 12:Layout of the 72-bar truss

    The last problem is a 200-bar planar truss which is schematized in Fig.13.The modulus of elasticityEof the material used in this structure equals 30,000 ksi while the material densityρequals 0.283 lb/in3.There are three independent load cases as follows:

    Table 3:Member groups for the 72-bar truss

    (LC1):1.0 kips acting along the positive x-direction at nodes (1), (6), (15), (20), (29), (34), (43),(48),(57),(62),and(71)(green arrows in Fig.13);

    (LC2):10 kips acting along the negative y-direction at nodes (1)–(6), (8), (10), (12), (14), (16)–(20),(22),(24),(26),(28)–(34),(36),(38),(40),(42)–(48),(50),(52),(54),(56)–(62),(64),(66),(68),and(70)–(75)(orange arrows in Fig.13);

    (LC3):both(LC1)and(LC2)acting together.

    There is no limitation of displacement in this problem.Stresses inside all members must be lower than±10 ksi.This structure has 29 groups of members as indicated in Table 4.It means this problem has 29 design variables A = {Ai,i=1, 2,..., 29} in whichAirepresents the cross-sectional area of members belonging theith group.Aiis limited in the range of[0.1,15]in2.

    4.2.2 Parameter Studies

    Obviously, the training dataset size and the number of weak classifiers strongly influence the accuracy of the AdaBoost model,thereby affecting the performance of the proposed method CaDE.In this section,a parametric study is carried out to investigate the influences of these parameters.The 25-bar truss structure described in Section 4.2.1 is taken as the optimization problem for the parametric study.

    Firstly,five cases are carried out in which the amount of training samples is set to 100,500,1000,1500,and 2000,respectively.In all cases,the number of weak classifiers used in the AdaBoost model is fixed asT=100.In the second,four more cases are conducted in which the number of weak classifiers is set to 10, 25, 50, and 200, respectively while the number of training samples is fixed as 1000 data points.There are a total of 9 cases.Other parameters are the same for all cases as follows:the mutation strategy“DE/target-to-best/1”;F=0.8;Cr=0.9;NP=50;andmax_iter=500.Each case is executed 30 times and the results are reported in Table 5.

    It can be found that all cases achieve high convergence when the optimal weights found in 30 independent runs are almost the same.However,there is an influence of the parameters on the number of structural analyses that must be conducted during the optimization process.When the training dataset size is fixed to 1000,the case of 10 weak classifiers requires the largest number of structural analyses.When the number of weak classifiers varies from 10 to 100,the number of structural analyses decreases linearly but from 100 to 200,the value ofnSAdoes not change much.When the number of weak classifiers is fixed to 100,the case of 1000 data samples has the fastest convergence speed.

    Overall, it is recognized that the pair of parameters:the number of weak classifierT=100 and the number of training samples=1000 is a proper selection.This set of parameters is then applied to all further experiments.

    Figure 13:Layout of the 200-bar truss

    Table 4:Member groups for the 200-bar truss

    Table 5:Influence of the parameters of the AdaBoost model on the performance of the CaDE

    4.2.3 Comparison with Regression-Based Surrogate Model

    As described in the‘Introduction’section,many previous studies have used regression techniques for surrogate modeling.In this section, a comparison is conducted to exhibit the advantage of the classification-based surrogate models.The performances of three optimization methods are evaluated including the original DE method(called the DE),the AdaBoost classifier-assisted DE method(called the CaDE),and the AdaBoost regression-assisted DE method(called the RaDE).

    The details of the DE and the CaDE are already introduced in Section 2.1 and Section 3.2.The RaDE has also two phases:the model building phase and the model employing phase.In the model building phase,the original DE algorithm is employed as usual and all solutions are exactly evaluated through structural analyses and the degree of constraint violation of each solution is saved in the database.Aftern_iter1 iterations,an AdaBoost regression model is trained with the obtained dataset with the aim of predicting the constraint violation degree of a new solution.Then,the trained model is employed instead of the constraint verification in the second stage.

    Five optimization problems mentioned in Section 4.2.1 are solved using three methods DE,CaDE,RaDE, respectively.The setting parameters for these methods are summarized in Table 6.Each problem is solved 30 independent times, and the statistical results including best, mean, worst, and standard deviation(SD)are reported in Table 7.The average number of structural analyses conducted during the optimization process (nSA) as well as the degree of constraint violation (cv) are also presented.

    Table 6:Parameters of the DE,CaDE and RaDE

    Based on the results in Table 7,some observations can be pointed out as follows.First of all,it can be seen that the optimal designs found by the CaDE are very similar to the DE but the CaDE needs fewer structural analyses than the DE.For the four first problems,the DE requires 25000 structural analyses and for the last problem,the DE requires 50000 structural analyses.However,the required number of structural analyses of the CaDE for five problems are 17311 times,16299 times,17599 times,18144 times,and 40683 times,respectively.

    Secondly, the required numbers of structural analyses of the RaDE for all problems are very small compared to the DE and the CaDE.This method conducts only 1000 times of exact structural analyses to generate the training data.However,due to the error between the structural analyses and the surrogate modeling,the optimal designs found by the RaDE are not good.Although these designs do not violate design constraints,the weights of these designs are much greater than the optimal designs found by the DE and the CaDE.

    Table 7:Comparison of optimal results of the DE,CaDE,and RaDE

    In general, using the regression-based surrogate modeling significantly reduces the number of structural analyses during the optimization process but the obtained results are not the global optimal.Otherwise,the classification-based surrogate modeling only reduces the number of structural analyses by about 25%,but the advantage of this method is that it retains the search performance of the original DE method.

    4.2.4 Comparison with Other DE Variants

    Furthermore,the results of these problems found by other DE variants are introduced for comparison.Four DE variants collected in the literature[40]include the composite differential evolution(CoDE), the self-adaptive control parameters differential evolution (JDE), the adaptive differential evolution with optimal external archive(JADE),and the self-adaptive differential evolution(SADE).The statistical results are introduced in Table 8.

    Table 8:Comparison of optimal results of the CaDE and other DE variants[40]

    For the 10-bar truss,the optimal design of the CaDE is similar to the CoDE and JADE and better than the JDE and the SADE.For the 17-bar truss, the optimal weight found by the CaDE is very similar to those of the DE variants.For the third problem of the 25-bar truss,the minimum weight of the CaDE(545.163 lb)is lower than those of other DE variants(545.555 lb).For the fourth problem of the 72-bar truss,the optimal weights found by the CaDE,the JDE,the JADE are the same(363.826 lb)and this value is slightly larger than the optimal weight obtained by the CoDE(363.824 lb).For the last problem,the weight of the 200-bar truss received from the CaDE is the smallest value(25490.653 lb),followed by the CoDE(25523.011 lb),the JDE(25579.000 lb),the JADE(25659.736 lb),and the SADE(25734.695 lb).Moreover, the smaller value of the standard deviation of the CaDE (111.277 lb) in comparison with the remaining algorithms(120.334 lb for the CoDE,191.068 lb for the JDE,301.116 for the JADE,and 518.116 lb for the SADE)indicates the reliability of the proposed method.

    In terms of the convergence speed,the CaDE exhibits its superior.The CaDE requires from 16299 to 40683 structural analyses while the remaining methods conduct 100000 structural analyses for finding the optimal results.It is apparent that the CaDE is more speedy than the CoDE, the JDE,the JADE,and the SADE.

    4.3 A Real-Size Transmission Tower

    In this section, the CaDE is employed to optimize the weight of a real-size transmission tower for showing its applicability in the practical design.This tower consists of 244 members which are classified into 32 groups as presented in Fig.14.Members belonging to the same group have the same cross-section which is chosen from 64 equal angle profiles from L35×3 to L200×25.It means this problem is a discrete optimization.The steel material used in this structure has mechanical properties as follows:the weight densityρ=7850 kg/m3;the modulus of elasticityE=210,000 N/mm2;the yield strengthFy=233.3 N/mm2.This structure is designed according to the AISC-ASD specification in which the tensile stress constraint is as follows:

    where:σtis the tensile stress of the member;[σ]tis the allowable tensile stress;Fyis the yield strength.

    The buckling stress constraint is as follows:

    where:σcis the compressive stress of the member;[σ]bis the allowable buckling stress;λ=KL/ris the maximum slenderness ratio of the member;Kis the effective length factor which equals 1 in this case;Lis the length of the member; andris the radius of gyration of the member’s cross-section;Ccis the slenderness ratio dividing the elastic and inelastic buckling regions which is determined using the following formula:

    Additionally,the maximum slenderness ratioλis limited to 200 for compression members while 300 for tension members.The tower is subjected to two independent load cases.Table 9 presents the force magnitudes of two applied load cases and the corresponding displacement limitation.

    Figure 14:Layout of the transmission tower

    Table 9:Load cases and the displacement limitations for the transmission tower

    This structure is optimized according to both the DE and the CaDE with the same parameters as follows:NP=50;F=0.8;Cr=0.9;max_iter=1000.The number of iterations for the first phase is set ton_iter1=300 which means 15000 samples have been collected for the training dataset.The statistical results obtained by the DE and the CaDE are presented in Table 10.The minimum weight of the tower found by the DE and the CaDE are the same(6420.652 kg)but other statistical results of the CaDE including mean, worst, and SD are smaller than those of the DE, indicating the stability of the proposed method.The number of structural analyses of the CaDE is only 24865 times,which is about 50%of the number of structural analyses performed by the DE.In other words,the CaDE omits about half of useless structural analyses.This is clearly shown in Fig.15.

    Table 10:Statistical results of the optimal designs found by the DE and the CaDE

    Figure 15:Comparison of the DE and the CaDE for the transmission tower problem

    4.4 Evaluation of the Effectiveness of the Proposed Method CaDE

    The average numbers of structural analyses(nSA)conducted in the six investigated examples as well as the reduction rates when using the AdaBoost classifier are listed in Table 11.Furthermore,the computing times of the DE and the CaDE are also reported.

    It is clearly seen that the application of the AdaBoost classifier into the optimization process helps to reduce the number of structural analyses.Specifically,the CaDE has reduced the number of structural analyses by up to 50.3%for the transmission tower problem.The minimum reduction rate is 18.6%for the 200-bar truss problem.

    Table 11:Average number of reduced structural analyses by using Adaboost classifier

    In contrast,the CaDE is slower than the DE in five benchmark truss problems when compared in terms of time.The CaDE is only faster than the DE in the transmission tower problem.This is because integrating the AdaBoost classifier into the DE increases the complexity of the algorithm.Moreover,the CaDE requires additional time for training and employing the AdaBoost model.However,for the transmission tower,this additional time is small in comparison with the computational time of fitness evaluation where involving structural analysis.Therefore the total computing time of the CaDE is lower than that of the DE in this case.In general, it can be concluded that the proposed method is effective with large-scale structures.

    5 Conclusions

    In this paper, the AdaBoost classifier-assisted Differential Evolution method, called CaDE, is proposed for optimizing the weight of truss structures.By integrating the AdaBoost classification technique into the DE algorithm,the unnecessary structural analyses are significantly reduced.More specifically,in the early generations,the original Differential Evolution algorithm is deployed but the data generated in this phase is collected to train an AdaBoost model with the aim of classifying the safety state of structures.In later generations,this model is used to detect and eliminate unpromising individuals.

    Through five optimization problems of truss structures,the CaDE method is proved to be a robust and reliable algorithm.In most cases, the CaDE method accomplishes the optimal results as good as or better than those of other algorithms in the literature but the numbers of structural analyses performed by the CaDE method are lower than those performed by the original DE algorithm.Besides advantages,the proposed method has also some drawbacks.Firstly,it is more complex than the original DE algorithm.Moreover, the proposed method requires more computational time for training and employing the machine learning model.Therefore,it is suitable for large-scale structures in which the structural analysis task is very time-consuming.

    In the future,the application of the CaDE method can be extended for other types of structures.Furthermore, using machine learning models to improve the performance of other optimization algorithms also forms an interesting research topic.

    Acknowledgement:The authors gratefully acknowledge the financial support provided by HUCE.Besides,the first author T.H.N.would also like to thank VINIF for the support.

    Funding Statement:This research is funded by Hanoi University of Civil Engineering (HUCE) in Project Code 35-2021/KHXD-TD.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    丝瓜视频免费看黄片| 男女边吃奶边做爰视频| 久久国产乱子免费精品| 国产极品天堂在线| 人妻夜夜爽99麻豆av| 在线天堂最新版资源| 欧美区成人在线视频| 国产精品1区2区在线观看.| 亚洲成人av在线免费| h日本视频在线播放| 三级经典国产精品| 国内精品一区二区在线观看| 亚洲,欧美,日韩| 汤姆久久久久久久影院中文字幕 | 国产精品久久久久久久电影| 亚洲欧美日韩东京热| 少妇熟女aⅴ在线视频| 国产精品1区2区在线观看.| 插逼视频在线观看| 亚洲欧美一区二区三区黑人 | 欧美+日韩+精品| 国产伦在线观看视频一区| 看免费成人av毛片| 成年女人看的毛片在线观看| 天天躁夜夜躁狠狠久久av| 日日啪夜夜撸| 国产精品av视频在线免费观看| 国产精品人妻久久久影院| 久久久精品欧美日韩精品| 亚洲aⅴ乱码一区二区在线播放| 亚洲四区av| 欧美xxxx性猛交bbbb| a级毛色黄片| 欧美高清性xxxxhd video| 亚洲久久久久久中文字幕| 国产精品麻豆人妻色哟哟久久 | 久久久久久久久久黄片| 麻豆国产97在线/欧美| or卡值多少钱| 亚洲av电影在线观看一区二区三区 | 你懂的网址亚洲精品在线观看| 91久久精品电影网| 精品人妻视频免费看| 久久精品久久久久久噜噜老黄| 色吧在线观看| 日本猛色少妇xxxxx猛交久久| 午夜亚洲福利在线播放| 2021少妇久久久久久久久久久| 韩国av在线不卡| 国产久久久一区二区三区| av在线老鸭窝| 国产午夜精品论理片| 亚洲最大成人中文| 高清午夜精品一区二区三区| 男人舔奶头视频| 久久久久久久久大av| 欧美性感艳星| 亚洲高清免费不卡视频| 国语对白做爰xxxⅹ性视频网站| 乱系列少妇在线播放| 欧美 日韩 精品 国产| 精品亚洲乱码少妇综合久久| 秋霞伦理黄片| 国内精品一区二区在线观看| 国产老妇女一区| 久久精品人妻少妇| 亚洲怡红院男人天堂| 亚洲av福利一区| 欧美激情在线99| 麻豆成人午夜福利视频| 久久久久久久国产电影| 国产高清三级在线| 小蜜桃在线观看免费完整版高清| 亚洲av二区三区四区| 国产成人精品久久久久久| 国产亚洲5aaaaa淫片| 在线观看一区二区三区| 97超视频在线观看视频| 噜噜噜噜噜久久久久久91| 卡戴珊不雅视频在线播放| 一本久久精品| 99热全是精品| 尤物成人国产欧美一区二区三区| 亚洲天堂国产精品一区在线| 国产伦精品一区二区三区四那| 久久精品熟女亚洲av麻豆精品 | av在线蜜桃| 只有这里有精品99| 亚洲av男天堂| 我要看日韩黄色一级片| 亚洲自偷自拍三级| 亚洲欧美日韩无卡精品| 肉色欧美久久久久久久蜜桃 | 亚洲精品国产av蜜桃| 午夜免费激情av| 99九九线精品视频在线观看视频| 亚洲av男天堂| 免费观看a级毛片全部| 美女脱内裤让男人舔精品视频| 天堂av国产一区二区熟女人妻| 色尼玛亚洲综合影院| 成人性生交大片免费视频hd| 日日摸夜夜添夜夜爱| 在线免费观看的www视频| 久久99蜜桃精品久久| 国产高清不卡午夜福利| 波野结衣二区三区在线| 2021天堂中文幕一二区在线观| 日本色播在线视频| 五月伊人婷婷丁香| 欧美精品国产亚洲| 免费观看性生交大片5| 简卡轻食公司| 亚洲成人一二三区av| av黄色大香蕉| 小蜜桃在线观看免费完整版高清| 岛国毛片在线播放| av网站免费在线观看视频 | 青青草视频在线视频观看| 午夜福利视频1000在线观看| 久久人人爽人人片av| 别揉我奶头 嗯啊视频| 免费看不卡的av| 精品熟女少妇av免费看| 五月伊人婷婷丁香| 亚洲欧洲日产国产| 午夜福利视频精品| 国产精品一二三区在线看| 最近最新中文字幕大全电影3| 免费av观看视频| 国产一级毛片七仙女欲春2| 亚洲精品日韩av片在线观看| 国产国拍精品亚洲av在线观看| 久久综合国产亚洲精品| 美女内射精品一级片tv| 亚洲成人一二三区av| 国产探花在线观看一区二区| 免费看光身美女| 国产乱人视频| www.色视频.com| 日本av手机在线免费观看| 亚洲精品日本国产第一区| 精品久久久久久成人av| 亚洲欧美成人精品一区二区| 久久久久久久久久人人人人人人| 2022亚洲国产成人精品| xxx大片免费视频| 国产高清有码在线观看视频| 人妻少妇偷人精品九色| 直男gayav资源| 欧美变态另类bdsm刘玥| 高清午夜精品一区二区三区| 国内精品美女久久久久久| 身体一侧抽搐| 免费观看a级毛片全部| 中文精品一卡2卡3卡4更新| 亚洲精品色激情综合| 国产黄a三级三级三级人| 日韩大片免费观看网站| 淫秽高清视频在线观看| 国产片特级美女逼逼视频| 蜜桃亚洲精品一区二区三区| 一级爰片在线观看| 网址你懂的国产日韩在线| 免费高清在线观看视频在线观看| 亚洲av免费在线观看| 久久韩国三级中文字幕| 秋霞在线观看毛片| 欧美97在线视频| 欧美最新免费一区二区三区| 汤姆久久久久久久影院中文字幕 | 国产亚洲av嫩草精品影院| 极品教师在线视频| 国产精品不卡视频一区二区| 欧美xxⅹ黑人| 久久99热这里只有精品18| 听说在线观看完整版免费高清| 国产淫片久久久久久久久| 亚洲天堂国产精品一区在线| 国产精品99久久久久久久久| 欧美精品一区二区大全| 日韩精品有码人妻一区| 久久久精品94久久精品| 亚洲国产最新在线播放| a级毛色黄片| 成人欧美大片| 国产精品一区二区在线观看99 | 国产国拍精品亚洲av在线观看| 少妇猛男粗大的猛烈进出视频 | 韩国av在线不卡| 内地一区二区视频在线| 免费看不卡的av| 成年女人看的毛片在线观看| av国产免费在线观看| 国产精品爽爽va在线观看网站| 免费观看的影片在线观看| 超碰97精品在线观看| 国产精品不卡视频一区二区| 国产成人精品久久久久久| 色吧在线观看| 亚洲高清免费不卡视频| 麻豆av噜噜一区二区三区| 成年女人在线观看亚洲视频 | 寂寞人妻少妇视频99o| 26uuu在线亚洲综合色| 欧美另类一区| 最近手机中文字幕大全| 晚上一个人看的免费电影| 国产淫语在线视频| 国产熟女欧美一区二区| av国产免费在线观看| 国产精品爽爽va在线观看网站| 亚洲va在线va天堂va国产| 亚洲美女搞黄在线观看| 人人妻人人看人人澡| av天堂中文字幕网| 哪个播放器可以免费观看大片| 日韩强制内射视频| 狂野欧美白嫩少妇大欣赏| 大又大粗又爽又黄少妇毛片口| 国产精品1区2区在线观看.| 如何舔出高潮| 久久久欧美国产精品| 午夜福利在线观看免费完整高清在| 日韩制服骚丝袜av| 国产免费视频播放在线视频 | 国产av在哪里看| 精品一区在线观看国产| 日韩 亚洲 欧美在线| 日韩视频在线欧美| 精品国产露脸久久av麻豆 | 卡戴珊不雅视频在线播放| 嘟嘟电影网在线观看| 日韩在线高清观看一区二区三区| 在线观看av片永久免费下载| 久久人人爽人人爽人人片va| 欧美xxxx黑人xx丫x性爽| 久久久久久久久久久丰满| 日韩国内少妇激情av| 蜜臀久久99精品久久宅男| 亚洲精品国产成人久久av| 人妻少妇偷人精品九色| 国产欧美日韩精品一区二区| 欧美三级亚洲精品| 亚洲av.av天堂| 欧美成人一区二区免费高清观看| av黄色大香蕉| av福利片在线观看| 国产不卡一卡二| 啦啦啦中文免费视频观看日本| 非洲黑人性xxxx精品又粗又长| 国产69精品久久久久777片| 欧美激情国产日韩精品一区| 亚洲成人精品中文字幕电影| 亚洲最大成人av| 国产精品久久久久久精品电影小说 | 国内少妇人妻偷人精品xxx网站| 天堂√8在线中文| 一级毛片我不卡| 午夜激情欧美在线| 精品久久久噜噜| 波野结衣二区三区在线| 在线免费观看的www视频| 男插女下体视频免费在线播放| 高清在线视频一区二区三区| 在线天堂最新版资源| 国产成人免费观看mmmm| videossex国产| av黄色大香蕉| 国产午夜福利久久久久久| 天堂√8在线中文| 国产中年淑女户外野战色| 亚洲,欧美,日韩| 在线观看一区二区三区| 国产在视频线精品| 18+在线观看网站| 美女主播在线视频| 久久久久九九精品影院| 久久久精品免费免费高清| 黄片wwwwww| 亚洲精品日本国产第一区| 少妇的逼水好多| 街头女战士在线观看网站| 欧美zozozo另类| 夜夜爽夜夜爽视频| 只有这里有精品99| 国产精品久久久久久av不卡| av在线老鸭窝| 22中文网久久字幕| 中国国产av一级| 蜜桃亚洲精品一区二区三区| 国产高清不卡午夜福利| 免费电影在线观看免费观看| 免费观看无遮挡的男女| 亚洲人成网站在线播| 国产探花极品一区二区| 亚洲精品日韩av片在线观看| 日日摸夜夜添夜夜爱| 免费黄网站久久成人精品| 亚洲成人av在线免费| 亚洲精品乱久久久久久| 春色校园在线视频观看| 日本一本二区三区精品| 国产熟女欧美一区二区| 日韩欧美三级三区| 中文乱码字字幕精品一区二区三区 | www.av在线官网国产| 国内揄拍国产精品人妻在线| 啦啦啦中文免费视频观看日本| 99热这里只有精品一区| 欧美最新免费一区二区三区| 高清日韩中文字幕在线| 在现免费观看毛片| 久久久久久久久中文| 国产成人精品久久久久久| 一级毛片 在线播放| 亚洲综合精品二区| 又黄又爽又刺激的免费视频.| 成人毛片a级毛片在线播放| 亚洲av电影在线观看一区二区三区 | 久久精品综合一区二区三区| 人妻系列 视频| 成人高潮视频无遮挡免费网站| 国产不卡一卡二| 一二三四中文在线观看免费高清| 亚洲精品乱码久久久v下载方式| 精品不卡国产一区二区三区| 国产精品一及| 中文乱码字字幕精品一区二区三区 | 成年女人看的毛片在线观看| 床上黄色一级片| 久久精品国产亚洲网站| 亚洲在久久综合| 久久99热这里只有精品18| 毛片女人毛片| 亚洲av中文字字幕乱码综合| 国产三级在线视频| 色播亚洲综合网| 亚洲av成人精品一区久久| 亚洲精品自拍成人| 日日干狠狠操夜夜爽| 日韩欧美国产在线观看| 成人毛片60女人毛片免费| 亚洲最大成人av| 免费黄网站久久成人精品| 国产成人精品久久久久久| av一本久久久久| 亚洲av不卡在线观看| 国产高清不卡午夜福利| 国产国拍精品亚洲av在线观看| 亚洲av一区综合| 99久久人妻综合| 精品人妻偷拍中文字幕| av国产久精品久网站免费入址| 精品久久久久久久人妻蜜臀av| 成人国产麻豆网| 最近的中文字幕免费完整| 白带黄色成豆腐渣| 欧美另类一区| av黄色大香蕉| 日产精品乱码卡一卡2卡三| 国产在视频线精品| 99re6热这里在线精品视频| 亚洲自偷自拍三级| 晚上一个人看的免费电影| 一区二区三区高清视频在线| 一二三四中文在线观看免费高清| 成年女人在线观看亚洲视频 | 亚洲av男天堂| 韩国高清视频一区二区三区| 成年女人看的毛片在线观看| 女人十人毛片免费观看3o分钟| 国产 一区 欧美 日韩| 哪个播放器可以免费观看大片| 一级毛片我不卡| 亚洲性久久影院| 一级片'在线观看视频| 亚洲在线自拍视频| 亚洲18禁久久av| 精品一区二区三卡| 淫秽高清视频在线观看| 特级一级黄色大片| 丝袜美腿在线中文| 看十八女毛片水多多多| 欧美成人一区二区免费高清观看| 亚洲第一区二区三区不卡| 亚洲四区av| 两个人的视频大全免费| 秋霞在线观看毛片| av免费观看日本| 午夜精品国产一区二区电影 | 久久人人爽人人爽人人片va| 久久这里有精品视频免费| av又黄又爽大尺度在线免费看| 国产综合懂色| 2018国产大陆天天弄谢| 国产成人午夜福利电影在线观看| 欧美性感艳星| 99久久九九国产精品国产免费| 国产成人a∨麻豆精品| av福利片在线观看| 99久国产av精品国产电影| 国产免费一级a男人的天堂| 成年人午夜在线观看视频 | 99久久精品国产国产毛片| 亚洲欧美精品自产自拍| 免费黄色在线免费观看| 亚洲精品日韩av片在线观看| 亚洲av在线观看美女高潮| 色综合亚洲欧美另类图片| 真实男女啪啪啪动态图| 亚洲国产欧美人成| 搞女人的毛片| 六月丁香七月| 国产精品一区二区性色av| 简卡轻食公司| 亚洲自拍偷在线| 亚洲av电影不卡..在线观看| 国产精品一及| 午夜免费观看性视频| 免费不卡的大黄色大毛片视频在线观看 | 日韩欧美精品v在线| 久久久久久久大尺度免费视频| 欧美日韩精品成人综合77777| 国产久久久一区二区三区| 亚洲一级一片aⅴ在线观看| 国产日韩欧美在线精品| 噜噜噜噜噜久久久久久91| 搞女人的毛片| 欧美日韩国产mv在线观看视频 | 女人久久www免费人成看片| 六月丁香七月| 两个人的视频大全免费| 天堂俺去俺来也www色官网 | 三级国产精品片| 国产真实伦视频高清在线观看| 熟妇人妻不卡中文字幕| 亚洲18禁久久av| 波多野结衣巨乳人妻| 国产综合精华液| 国产白丝娇喘喷水9色精品| 午夜免费激情av| 男的添女的下面高潮视频| 青春草国产在线视频| 美女大奶头视频| 熟妇人妻不卡中文字幕| xxx大片免费视频| 欧美高清成人免费视频www| 一个人免费在线观看电影| 国产白丝娇喘喷水9色精品| 99久久精品国产国产毛片| 日韩av不卡免费在线播放| 女人被狂操c到高潮| 热99在线观看视频| 国产成人精品一,二区| 精品一区在线观看国产| 黄色一级大片看看| 91午夜精品亚洲一区二区三区| 麻豆乱淫一区二区| 亚洲美女视频黄频| 欧美97在线视频| 青春草亚洲视频在线观看| www.av在线官网国产| 午夜免费激情av| 男女啪啪激烈高潮av片| 国产精品蜜桃在线观看| 亚洲一级一片aⅴ在线观看| 成人午夜精彩视频在线观看| av在线播放精品| 全区人妻精品视频| 99热全是精品| 国产精品av视频在线免费观看| 国产精品麻豆人妻色哟哟久久 | 日韩成人伦理影院| av在线天堂中文字幕| 欧美激情久久久久久爽电影| 亚洲最大成人手机在线| 人人妻人人澡欧美一区二区| 九九爱精品视频在线观看| 五月玫瑰六月丁香| 亚洲,欧美,日韩| 国模一区二区三区四区视频| 91午夜精品亚洲一区二区三区| 晚上一个人看的免费电影| 日韩大片免费观看网站| 亚洲欧美成人精品一区二区| 18禁在线无遮挡免费观看视频| 纵有疾风起免费观看全集完整版 | 久久久久久久久久人人人人人人| 国产美女午夜福利| 欧美日韩在线观看h| 国产在视频线在精品| 精品久久久久久久久av| 国产探花极品一区二区| 大香蕉97超碰在线| 免费看a级黄色片| av线在线观看网站| 一级毛片电影观看| 一个人看的www免费观看视频| 亚洲国产精品成人久久小说| 婷婷色综合大香蕉| 国产色婷婷99| 亚洲精品久久午夜乱码| 欧美xxⅹ黑人| 国产精品伦人一区二区| 亚洲精品乱久久久久久| 国产成人一区二区在线| 美女cb高潮喷水在线观看| 国产av在哪里看| 精品亚洲乱码少妇综合久久| 91久久精品国产一区二区三区| 久久久久久久久久久免费av| 熟女人妻精品中文字幕| 又大又黄又爽视频免费| 国产伦精品一区二区三区四那| 亚洲欧美日韩卡通动漫| 成人毛片60女人毛片免费| 欧美3d第一页| 国产成年人精品一区二区| 秋霞伦理黄片| 男女视频在线观看网站免费| 久久热精品热| 别揉我奶头 嗯啊视频| 欧美另类一区| 夫妻性生交免费视频一级片| 亚洲精品第二区| 久久99热这里只有精品18| 波多野结衣巨乳人妻| 一个人观看的视频www高清免费观看| 成人毛片a级毛片在线播放| 亚洲av不卡在线观看| 天堂√8在线中文| 少妇熟女aⅴ在线视频| 99久国产av精品国产电影| 亚洲欧美日韩无卡精品| 免费少妇av软件| 一个人免费在线观看电影| 国产成人精品久久久久久| 一级av片app| 欧美精品国产亚洲| 岛国毛片在线播放| 亚洲欧美一区二区三区国产| 99久久精品热视频| 久久热精品热| 高清毛片免费看| 久久久精品94久久精品| 免费看光身美女| 国产淫片久久久久久久久| 26uuu在线亚洲综合色| 成人欧美大片| 亚洲高清免费不卡视频| 国产极品天堂在线| 十八禁国产超污无遮挡网站| 观看美女的网站| 麻豆久久精品国产亚洲av| 免费在线观看成人毛片| 亚洲一级一片aⅴ在线观看| 特级一级黄色大片| av女优亚洲男人天堂| 久久精品国产自在天天线| 99久久精品一区二区三区| 国产精品久久久久久av不卡| 日日啪夜夜撸| 十八禁国产超污无遮挡网站| 久久99热这里只频精品6学生| 大话2 男鬼变身卡| 久久久久久久国产电影| 激情五月婷婷亚洲| 高清午夜精品一区二区三区| 嘟嘟电影网在线观看| 一边亲一边摸免费视频| 亚洲成色77777| 麻豆精品久久久久久蜜桃| 黑人高潮一二区| 欧美激情久久久久久爽电影| 最近的中文字幕免费完整| 国产精品一及| 国产一区二区三区综合在线观看 | 丰满少妇做爰视频| 亚洲乱码一区二区免费版| 久久精品国产鲁丝片午夜精品| 噜噜噜噜噜久久久久久91| av.在线天堂| 中文字幕亚洲精品专区| 91久久精品国产一区二区三区| 高清午夜精品一区二区三区| 国产v大片淫在线免费观看| 亚洲国产欧美人成| 九九在线视频观看精品| 久久草成人影院| 欧美另类一区| 亚洲国产高清在线一区二区三| 欧美 日韩 精品 国产| 色哟哟·www| 在线a可以看的网站| 欧美3d第一页| 日本猛色少妇xxxxx猛交久久| 欧美激情久久久久久爽电影| 日本一二三区视频观看| 婷婷色麻豆天堂久久| 99久久人妻综合| 少妇裸体淫交视频免费看高清| 成年女人看的毛片在线观看| 韩国av在线不卡| 国产精品av视频在线免费观看| 欧美变态另类bdsm刘玥| 高清欧美精品videossex| 黄片wwwwww| 91精品一卡2卡3卡4卡| 国产老妇伦熟女老妇高清| 菩萨蛮人人尽说江南好唐韦庄| 欧美激情国产日韩精品一区| 水蜜桃什么品种好| 97超视频在线观看视频| 白带黄色成豆腐渣|