• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Efficient sampling for decision making in materials discovery*

    2021-05-24 02:24:42YuanTian田原TurabLookmanandDezhenXue薛德禎
    Chinese Physics B 2021年5期
    關(guān)鍵詞:田原

    Yuan Tian(田原), Turab Lookman, and Dezhen Xue(薛德禎),?

    1State Key Laboratory for Mechanical Behavior of Materials,Xi’an Jiaotong University,Xi’an 710049,China

    2Los Alamos National Laboratory,Los Alamos,NM 87545,USA

    Keywords: sampling methods,active learning,decision making,material design,Bayesian optimization

    1. Introduction

    Material research and development have so far largely been based on expert knowledge and “trial-and-error”. This has the drawback of relatively low efficiency and high costs.[1]With improvements in measurement capabilities and advances in scientific computing, materials data is being accumulated at unprecedented levels.[2]Over the last ten years, the information sciences have played a pivotal role in the analysis and prediction from data. Together with experiment, theory, and computer simulations,[3–5]we are at the threshold of a new age in which we will be able to obtain relationships between composition, structure, processing, and performance faster, more accurately, and more economically than ever before.[6–10]In addition,we will be able to employ the tools of decision making to guide the discovery of targeted materials with the fewest experiments to minimize the costs and labor.[11–21]

    Materials science is characterized in the main with data that is often expensive to obtain as experiments are not easy,and it is accompanied by a materials space that is vast and complex. Hence,often a relatively small size of data is available for training and we have to make predictions in a high dimensional materials design space.[22–24]Machine learning models trained on such small data sets are then usually accompanied with large uncertainties,which makes the decision making process challenging.[12–21,25–29]Using methods to improve decision outcomes and effectiveness is essential in guiding and accelerating the discovery of materials with targeted properties.[15–21,26,28,30]

    A supervised learning approach requires a substantial amount of labeled data to train the model adequately.[31]However,in materials science we usually have a large pool of unlabeled data. This is where a subfield of machine learning,known as active learning,[32–34]becomes important. Active learning depends on taking relatively few special labeled data points, those labeled by an expert, and training a model on those data. The claim then is that active learning can achieve a similar or better performance using fewer data than a supervised model does.[28,37–40]Thus,active learning endeavors to select the most“informative”data samples in order to successively improve the model in an iterative loop.[35,36,41,42,42–44]It involves a continuous learning process, in contrast to passive learning which relies on making predictions on the available training data. It is also called “query learning” as it depends on selecting or querying an oracle or experiment for the next label,and is related to“optimal experimental design”[12,45]or“sequential design”.[46,47]Bayesian optimization and adaptive design are methods following the similar strategy, which require goal-directed iterative feedbacks.[21,26,28]

    The different ways in which a learner can make queries include (a) query synthesis, (b) stream-based selective sampling,and(c)pool-based sampling.[32]In query synthesis,the learner learns by making membership queries[48–50]from any unlabeled data in the input space, such as the feature space,and queries the oracle for the label. Though quite reasonable on certain problems,[51]query synthesis can give rise to problematic choices, for example,ambiguous images when learning to discriminate characters. In stream-based sampling,[52]typically each unlabeled data is drawn continuously from an underlying distribution and the learner decides whether to query it or to discard it. This latter decision can be made based on defining a utility function or information measure so that samples with higher utility are more likely to be queried.Whereas one data point is sampled sequentially from a streaming data source in stream-based learning, pool-based active learning[53,54]considers all of the unlabeled data pooled together before making the selection using a utility as a means to sample. It is the most common learning situation for active learning,especially in applications to materials problems.

    Figure 1 shows an active learning loop with two aspects,the learner or model prediction and the query. Starting from a “small data” set, there is an inference to establish a model with uncertainties. The results of the prediction or probability estimates are applied to a pool of unlabeled instances or data. A number of query methods can then be employed for deciding on the next selection. We will review below several pool-based sampling methods including committee-based strategies,large margin-based strategies,posterior probabilitybased strategies,sampling strategies based on directed exploration, and mean objective cost of uncertainty. The sampling then guides the next measurement(oracle)to label the data.

    Fig.1. Active learning loop emphasize on the query section.

    In this review,we outline in Subsection 2.1 the motivation underlying commonly used sampling methods,which use uncertainties in model predictions to make the next queries. We also discuss other uncertainty based sampling strategies,such as methods based on directed exploration and mean objective cost of uncertainties in Subsections 2.2 and 2.3, to guide the choice of the next samples to test. We conclude in Section 3 with a number of examples from materials science related to classification and regression applications,which show how active learning can be applied in practice.

    2. Sampling strategies

    Selecting specific data samples for labeling from a pool of unlabeled data depends on the sampling strategy. Active learning uses these strategies in the form of utility or acquisition functions learned by constructing a model from data,that is the expected utilities. The sample with the largest utility is then chosen to be labeled. The sampling strategies includes committee-based strategies,large margin-based strategies,and posterior probability-based strategies, etc. We will elaborate these sampling strategies.

    2.1. Commonly used sampling strategies

    Committee-based strategies Here we use the predictions from several models to guide us on which data samples to label. The set of models is known as a committee[55,56]so that k different models will provide k labels, and the data sample chosen is the one giving maximum disagreement in predictions. The distinct models for the same dataset can be with different hyperparameters or different algorithms such as support vector machines (SVM),[57]decision trees, logistic regression, etc. The disagreement itself in predictions obtained from the committee can be measured using the concept of entropy,which is an aspect of uncertainty sampling.[58]Thus,if V(yi) indicates the number of votes for the label yi, then the chosen sample is x*that maximizes the entropy function given by[59]

    where C is a constant.

    The KL-divergence is a slight modification of the entropy definition and can also be utilized to represent the difference between two probability distributions p(y)and q(y). It allows us to evaluate how much information is lost when we approximate one distribution by another. The KL-divergence is defined by[60,61]

    The data sample which produces most disagreement across the predictions is labelled accordingly using, say, the above methods,and for regression-based models the variance can be used as a measure of disagreement.[62]Thus,high variance in the predictions represents samples likely to be most informative. We can also use bootstrap or bagging[63]in conjunction with entropy or variance approaches. The k training sets are created from the original data set by drawing samples randomly with replacement to create k predictions from k models. Another approach to construct a committee of models is to use the feature space so that each model works on the same data set but with different sets of features. This can potentially provide a greater degree of divergence in the predictions of the models.

    Large margin-based strategies Margin-based strategies are especially relevant for classification problems. The model uncertainty of classifiers such as the margin-based SVM can be represented by the distance to the separating hyperplane.[64]Evaluating the distance to the separating hyperplane is the straightforward way to measure the model uncertainty on unknown sample. The support vectors of the SVM define the decision boundary and could be considered as most informative samples or those with greatest uncertainty. While a sample away from the decision boundary considered with a high confidence of class assignment will be thus not interesting for future sampling. Thus,the margin sampling method considers the data sample for which the prediction on one of the classes is most uncertain, corresponding to those with minimum distance from the hyperplane. The informative data samples fall within the margin and the samples or support vectors chosen,which are given by

    where f(xi,ω)represents the distance between the data sample and the hyperplane for class ω. One single data sample is chosen for querying the oracle per iteration, so there is computational overhead. In contrast,the multi-class level strategy considers a multi-class setting. For a data sample,the classes with the largest and second-largest distances from the margin are noted by identifying the distances from the hyperplane for each class of each data sample. The multi-class level uncertainty is then defined as the two largest distances corresponding to two different classes. Thus,[65]

    The goal is to minimize this difference so that the most uncertain and informative unlabeled data sample can be chosen.

    Posterior probability-based strategies The posterior probability distribution provides a model’s ability and uncertainty in assigning a particular class to a data sample and can also be queried to determine whether an unlabeled data sample should be labeled or not. Such a strategy can be used with any model capable of predicting output probabilities. A well known approach is the least confidence which selects the next sample with least certain in its prediction or class assignment.Thus,if Pθ(?y|x)is the highest posterior probability for label ?y predicted by model θ,and maximizing(1 ?Pθ(?y|x))gives us the data sample with the highest uncertainty. That is,[66]

    The strategy can be extended to multiple classes in which we consider the two most probable classes. One has the highest probability of ?y1and the other has the second-highest probability of ?y2. Both probabilities are obtained using model θ. We can minimize the difference between the two most probable classes, as it is indicative of the uncertainty in predicting the unexplored data sample by the model. Therefore the smallest difference represents the most uncertain and informative sample,which is given by[67]

    Such a sampling method sometimes is also referred as margin sampling, similar with the large margin-based strategy mentioned above. The difference is that this algorithm uses the posterior probabilities,while the other is based on the margin-based models. Going beyond one or two most probable classes for sampling,requires us to consider information lying in the remaining class probability distributions as it is unused. This is where the entropy measure is useful for sampling. The entropy is a measure of disorder or impurity in a system, which can be also used as a measure of model uncertainty. Hence, higher entropy values indicate high model uncertainty related to class membership. The data sample for which the model is least certain or most uncertain about the class assigned can be recommended by maximizing the following entropy function:[68]

    We have seen the importance of several measures of uncertainty in the general sampling approaches outlined above.Entropy, margin, variance, and least confidence all provide definitions of uncertainties within the appropriate modeling context. Here we will consider a number of other sampling approaches recently employed in materials science that use uncertainties intimately to sample new data points for measurements.

    2.2. Sampling strategies based on directed exploration

    Sampling in an active learning loop often demands exploring the whole compound space adequately to find the optimal targets. Different degrees of exploration,i.e.,how much exploration needs to be performed,are determined by the costs of collecting new information and the value associated with that information as experiments in materials science are often costly. It is thus important to choose an appropriate degree of exploration so that the most important states can be efficiently sampled,especially when there is much unknown in the whole space.

    Two types of exploration techniques are known as the undirected one and the directed one. The undirected exploration selects candidates randomly from some distribution,such as uniform distribution, semi-uniform distribution, and Boltzmann distribution. In the uniform-distribution case,random exploration is taken in which costs or rewards are not taken into account. If the distributions are other than uniform, the same treatment will be implemented as the actions are random. Whitehead[74]proved that random walk exploration leads to the learning time scaling exponentially with the size of the state space under certain conditions.

    The other exploration technique is the directed exploration.[73]This kind of exploration uses some criteria or knowledge to guide the exploration. It helps to make the right decision to explore the environment best, which aims at maximize the improvement over time. However, it is impossible to know how much improvement of the candidate’s performance unless it has been measured. Thus, all directed exploration techniques are designed to optimize knowledge gain. The exploration may be achieved by choosing states based on frequency of occurrence(counter based exploration),high predicted uncertainty (maximum variance) or functions including various degrees of exploitation. Various trade-off methods,such as efficient global optimization(EGO),knowledge gradient(KG),based on Bayesian optimization also fall in this broad category.

    Maximum variance If the purpose is to predict the continuous values such as hardness, strength, temperature, etc.,the variability at each point can be computed through bootstrap methods[75]and the Jackknife method, etc.[76–79]The uncertainty is thus given by

    Maximum variance samples the point with the largest uncertainty in the whole space. If the deviation of the true value from the predicted value is large, the variance can be also large. The maximum variance chooses a point with the largest variance which can reduce the total deviation after measuring that point. It can be employed in many materials problems to optimize a material property curve/surface,such as the potential energy surface for Ar-SH[80]and the phase boundary.[27]

    Pure exploitation method This chooses the point corresponding to the‘best’model prediction for the next selection

    Similarly,pure exploration corresponds to the the choice with the maximum variance.

    Exploration–exploitation trade-off methods Both the predicted value and the variance serve as input to these methods.

    (a) Efficient global optimization (EGO). The purpose is to maximize the expected improvement over the best observed so far,ymax,[81,82]

    where s is the uncertainty associated with the mean value ˉμ*of the model prediction,φ(·)and Φ(·)are the probability density function and cumulative distribution function of normal distribution,respectively.

    (b) Knowledge gradient (KG) is designed to measure the alternative that produces the greatest value from a single observation[83–86]

    2.3. Mean objective cost of uncertainty

    The emphasis of mean objective cost of uncertainty(MOCU) is on minimizing the uncertainty in the objective function itself rather than on reducing the overall uncertainties or entropy to track information in a system.[69,70]A prior distribution is assumed,which is then updated to form a posterior after the outcome of an experiment is known. MOCU tracks how a system loses its performance because of the presence of uncertainty and the next experiment is chosen to reduce MOCU, i.e., that which reduces the variance in the posterior distribution the most.[71]

    For a cost function f(x), let us define xrobustas the point with the expected value of f(x) over the unknown parameters,θ. That is,xrobust=argmaxxEθf(wàn)(x)is the best or‘a(chǎn)verage’ result we can obtain given that θ is unknown. It is not the optimal because we do not know θ, but if θ was known,then the optimal would be given by x+=argmaxxfθ(x) for that value of θ. The difference between the optimal value f(x+) and the value associated with the robust selection,fθ(x+)?fθ(xrobust), is the loss we incur due to uncertainty for each value of θ. This difference is known as the objective cost of uncertainty (OCU) and its expectation over θ defines MOCU.That is,we define the expected MOCU as

    The next experiment is then selected by considering all possible outcomes of a given experiment and choosing that which reduces MOCU the most. The application of MOCU within a modeling framework for shape memory alloys is discussed in the study of Dehghannasiri et al.[72]

    3. Accelerating materials discovery

    3.1. Model prediction

    For classification problems, logistic regression, decision tree,random forest,gradient-boosted tree,support vector machine (SVM), and naive Bayes can be used to construct a model. The random forest and gradient-boosted tree are ensemble models making use of predictions on an ensemble of models to make robust predictions. Decision trees are usually the base learners and the ensemble is over many such trees.The SVM performs non-linear classification using a kernel function to map the problem to a high-dimensional feature space. It uses a hyper-plane to fit the data. Naive Bayes is a probabilistic classifier based on Bayes’theorem and provides the probability for the next sampling.

    Regression models include linear, polynomial, support vector regression using a kernel function, and the Gaussian process based on a Gaussian kernel for the covariance. Polynomial regressor is usually fit by minimizing the variance of the unbiased estimators of the coefficients.

    3.2. Case studies

    3.2.1. Posterior probability-based strategies

    Example 1: phase diagram construction We review an example where uncertainty sampling methods are employed to efficiently construct a phase diagram.[66]The active learning loop(Fig.2)shows how starting with a phase diagram with few labeled points, a new candidate may be sampled to optimize the initial phase diagram. The probability distributions P(p|x)of the phases for each point x can be obtained through semi-supervised learning models including label propagation(LP)and label spreading(LS),where p is the label of phases.In LP,the labels of the measured points are fixed. LP provides the probability of a phase p at x by calculating that of reaching the phase p first via random walk from x. The labels of measured points can be changed according to different environment in LS so that LS is effective when the label noise is large.Each unmeasured point x on the phase diagram is thus labeled by choosing the phase with the highest probability. The uncertainty score is calculated via different sampling approaches.Three uncertainty sampling (US) methods mentioned above are adopted in this work: the least confident(LC)method,the margin sampling(MS),and the entropy-based approach(EA).The unmeasured points located closest to the margin of multiple phases are of higher uncertainty than those further away from the margin. The next candidate is determined from the unmeasured points with the highest uncertainty score.

    Uncertainty sampling is compared with random sampling(RS)to construct the phase diagrams of H2O under lower pressure (H2O-L), H2O under higher pressure (H2O-H), and the ternary phase diagram of glass-ceramic glazes of SiO2,Al2O3,and MgO(SiO2–Al2O3–MgO).LC(LP+LC)is used to evaluate the uncertainty score. Begin with the same nine data points,after several given iterations,the samplings follows different trajectories using different sampling methods,shown in Fig. 3. Figures 3(d)–3(f) are the result of sampled points by uncertainty sampling. These sampled points are apparently distributed near the phase boundaries, which represent well the phase boundary locations. However, it is found that random sampling points hardly capture the margin between the two classes if one checks the same number of points as US does. Moreover,if there are small areas of certain phases,it is difficult for RS to sample points in this area,and this area will be easily ignored. For example,no point is sampled in region III of the H2O-H phase diagram in the RS approach. Efficient estimation of this region is presented in Fig.3(e)using US.

    Fig. 2. The schematic of phase diagram estimation and optimization process based on the uncertainty sampling methods. Reproduced with permission from Ref.[66].

    Fig. 3. Phase diagrams for (a) H2O-L, (b) H2O-H, (c) SiO2–Al2O3–MgO and the corresponding results of these phase diagrams utilizing(d)–(f)uncertainty sampling and(g)–(i)RS approaches. Reproduced with permission from Ref.[66].

    Fig.4. Performances of different uncertainty sampling approaches on the phase diagram construction. Reproduced with permission from Ref.[66].

    The performances of various uncertainty sampling methods are quantitatively shown in Fig.4.The results indicate that the LP method is better suited than the LS method. The EA is found not to be useful. If the number of phases is small, MS is better suited, whereas LC is powerful when many phases co-exist in a phase diagram. Thus, efficient selection can be realized using LC to construct complicated phase diagrams.Uncertainty sampling can decrease the number of sampling points to 20% and still construct an accurate phase diagram.Furthermore, the US approach can find an undetected new phase rapidly and a smaller number of initial sampling points are sufficient to obtain an accurate phase diagram.

    3.2.2. Sampling strategies based on directed exploration

    3.2.2.1. Single property optimization

    Example 2: search for very low thermal hysteresis(ΔT) NiTi-based shape memory alloys. To pursue a desired property, alloys with multi-component are usually considered. The increase in the number of elements and the variability of the elemental concentrations lead to an exponentially increased search space. It is thus a challenge to search for a new material in such a huge unexplored space.The work in Ref. [28] demonstrates an active learning loop consisting of 1) data collection, 2) inference model, 3) experimental design, and 4) feedback to realize the optimization of materials property, shown in Fig. 5. After training 22 measured data synthesized in the laboratory by a support vector regression/Gaussian process regression model,properties and uncertainties of the candidates in the large virtual space Ni50?x?y?zTi50CuxFeyPdzof 797482 potential alloys can be predicted. To guide the next measurement,different utility functions such as KG/EGO, which combine exploration and exploitation and the greedy pure exploitation method Min are used. We start from randomly selected samples and combine different models with different utility functions to find the optimal alloy. For robust results,we repeat this process 2000 times to investigate the performances of several regressors: utility function combinations as a function of the size of the data. SVRrbf: KG is found to outperform other regressor: utility function combination on the training set,shown in Fig.6.

    Fig. 5. The adaptive design loop based on the model predictions and uncertainties is demonstrated to guide the experiments efficiently. Reproduced with permission from Ref.[28].

    Fig. 6. The performances of various combinations of regressor and selector on the NiTi SMA training data set. Reproduced with permission from Ref.[28].

    The predicted value and its associated uncertainty for thermal hysteresis (ΔT) of all the data in the search space are provided using 1000 bootstrap samples with the SVRrbfbase learner. This is followed by KG to explore the whole space and focus on a local area with the apparent global minima. Ti50.0Ni46.7Cu0.8Fe2.3Pd0.2possessing the smallest ΔT(1.84 K) was discovered in the sixth iterations of our loop.Our design framework thus provides the opportunity for efficient decision making to accelerates the process of finding materials with desired properties.

    Example 3:finding high phase transition temperature NiTi-based shape memory alloys.We demonstrate a statistical learning approach encoded an active learning loop to predict martensitic transformation temperatures in this work.[29]Material descriptors related to the physical properties and the microstructure such as chemical bonding, atomic radii, etc.are employed in the model to capture the possible mechanisms underlying dependence of the transformation temperature, which enable us to expand the method to other alloys. Pauling electronegativity (en), valence electron numbers (ven), Waber–Cromer’s pseudopotential radii (dor) are chose as the final features through Pearson map followed by the best subset selection methods. We compare different machine learning models built by the three preselected features,including linear model, polynomial model, and support vector regression model (rbf, linear, poly kernel). The performance of the machine learning inference models on the training dataset is shown in Fig.7,and the best performer SVR with rbf kernel is decided as the model in the adaptive design framework. Based on the predicted mean and standard deviation provided by the bootstrap results based on the SVR model,decision making can be completed via utility functions. The designed loop chooses the best candidate for synthesis and characterization. We show that iteratively learning and improving the statistical model can accelerate the search for SMAs with targeted transformation temperatures. The Ti50Ni25Pd25with the highest transition temperature of 182.89°C was discovered in the Ti50(Ni50?x?y?zCuxFeyPdz)alloy system of totally 1652470 possibilities.

    In addition, there are some domain knowledge in material science. Combine the domain knowledge with the active learning loop will make the sampling much more efficient.Examples include accelerated search for BaTiO3-based piezoelectrics with vertical morphotropic phase boundary using Bayesian learning.[26]With the knowledge of the vertical MPB providing temperature-independent d33piezoelectric property for the compounds here, the solid solution (Ba0.5Ca0.5)TiO3-Ba(Ti0.7Zr0.3)O3is predicted and validated with piezoelectric properties that show good temperature reliability. Another example is accelerated search for BaTiO3-based ceramics with large energy storage at low fields.[37]In this work, the size of virtual space is shrunk from ~9 million to seven hundred thousand based on the domain knowledge, combining with machine learning, cooperating with sampling method EGO,we find the compound (Ba0.86Ca0.14)(Ti0.79Zr0.11Hf0.10)O3with the largest energy storage density 73 mJ·cm?3at a field of 20 kV·cm?1in only two iterations.

    Fig. 7. The performances of the machine learning inference models: (a) linear regression (LIN); (b) polynomial regression (POLY); (c) support vector regression with radial basis function kernel (SVR.rbf); (d) support vector regression with linear kernel (SVR.lin); and (e) support vector regression with polynomial kernel(SVR.poly). Reproduced with permission from Ref.[29].

    3.2.2.2. Multi-objective optimization

    Example 4: design of a new Mg alloy with high strength and ductility. In the process of developing new materials with optimized performance, we often desire that the material has more than one excellent property, and even expect these properties to improve simultaneously during the optimization process. However,the relationship of many properties requires a balance as not all can be equally improved. The most typical example is the strength and ductility of metals.Therefore, in order to gain the new materials with multiple excellent properties with as few experiments as possible, an efficient decision making method is needed as a guide.

    In this work,[87]we applied the machine learning assisted multi-objective optimization for materials processing parameters aiming at enhancing the strength and ductility of an as-cast ZE62 (Mg 6 wt.% Zn–2 wt.% RE (Y, Gd, Ce, Nd)) magnesium alloy. All the steps including machine learning prediction of multiple objectives,efficient sampling,and experimental feedback are enclosed in an active learning strategy to perform multi-objective optimization of the as-cast ZE62 magnesium alloy. The samplings are completed through scalarizing a set of objectives into a single objective. Figure 8 shows the schematic of our design strategy,illustrating how to search for the best solution to meet the trade-off optimization requirements. The goal is to push the new discovery materials to the target(yti).

    The famous method solving the multi-objective problem is to find the Pareto front. In the first approach,we decide the Pareto front firstly, then aim to minimize the included angle θpof two vectors(ωtand ωp)in Fig.8. Between them,vector ωtis defined from the origin towards the target. Vector ωpis from the origin towards the point in the Pareto front of the virtual space. The next candidate decided is calculated by the formula x=argminθp,where θpis given by

    The option x=argmin(δj)will then be chosen as the next experiment to be performed. These methods both transform the two-objective optimization problem into a single-objective optimization problem.

    Fig.8.The schematic of our design strategy illustrating how to search for the best solution to meet the trade-off optimization requirements. Reproduced with permission from Ref.[87].

    We utilized the two strategies to sample the the heat treatment conditions of alloy ZE62 among the large parameter space. According to the recommendation, perform the corresponding heat treatments and characterize the strength and ductility via the stress–strain curves. The active learning loop runs till the requirements are met. As a result, the strength and ductility of the alloy are increased by 27%and 13.5%respectively with only four new experiments. This work offers two sampling recipes to solve the multi-objective optimization problems in designing processing parameters of materials.

    3.2.2.3. Material property curve/surface optimization

    Example 5: efficient estimation of material property curves and surfaces. Material properties are always affected by variables such as micro-structure parameters, compositions, the heat treatment conditions as well as the ambient environments. A material property curve or surface exhibits property changes and makes convenient the production process. However,with each new component or materials parameter,the space of candidate experiments grows exponentially.Especially when the target curve/surface is very complex with critical points, a large number of steps will be taken to determine such a curve/surface through experiments or calculations. Thus,learning from existing data complied by sampling the most informative options from the search space will realize the construction of the curve/surface with the least cost. In this work,[80]we employ 5 kinds of sampling methods based on exploration compared with random sampling in different feature dimensions, including 1) one-dimensional cases: the fatigue life curve for 304L stainless steel (SS304L) and the liquidus line of the Fe–C phase diagram; 2) multi-dimension cases: the intermolecular potential of Ar-SH and the Curie temperature of BaTiO3ferroelectrics modeled in 4D.

    Figure 9 is the optimization result for the fatigue life curve of SS304L steel. We start from a random training data set of n=5 points and calculate the deviation from the prediction curve to the true one. Mean absolute error(MAE)and maximum absolute error(Max.AE)as a function of the number of new measurements for the different sampling methods are tracked. Figure 9 plots the average values and 95% confidence intervals for MAE and Max.AE for the different utility functions over 100 trials. Maximum variance(Max-v)and B.EGO aiming at searching for the option with the maximum of the variability in the function perform well. These two are able to converge in relatively few iterations, followed by random, which also converges but with more iterations. The trade-off methods EGO and KG decrease the error quickly in the first three iterations,but then relax slowly and nevertheless also converge after a large amount of experiments.The greedy,pure exploitation Min-u shows very little relaxation after a few iterations.

    Fig. 9. The performance of five utility functions compared with random selection in optimizing the fatigue life curve for SS304L steel. (a) and (b)Mean absolute error and maximum absolute error during the optimization process. Reproduced with permission from Ref.[80].

    Here, the 1D case of the fatigue life curve for SS304L steel is reviewed and the performances of different sampling strategies in other cases are shown in Ref.[80]. In general,directed exploration via maximum variance performs better than the others in mapping the property curve. Following Maxv, the variability utility function B.EGO based on bootstrap samples is sometimes also a good performer. However,under some circumstances,the trade-off methods employing various degrees of exploitation can perform at least as well as maximum variance(Max-v),even better than Max-v in a few cases.The choice of the sampling methods is sensitive to many factors such as the distribution of the training and subsequently acquired data,the model performance,the noise as well as the budget,which determines the number of iterations allowed.

    4. Conclusion

    We have provided an overview of sampling strategies,including uncertainty sampling as well as those based on directed exploration of the search space. Trade-off methods,which balance exploration with exploitation functions that use the best model outcomes, belong to the latter class and have been used widely in sequential decision making. Efficient sampling guides the selection of the optimal next experiment and is a key component in minimizing the number of experiments and hence the costs required. We have shown how several materials problems can be addressed by employing sampling strategies, and one of our conclusions is that maximum variance is a robust strategy across several examples we have investigated. There is a need for further numerical and analytical studies to distill guiding principles for what strategies are best suited for different types of materials data.

    猜你喜歡
    田原
    Temperature-free mass tracking of a levitated nanoparticle
    河流
    田原發(fā)表全新EP《2080》第二首單曲《WHY 2080》
    青年歌聲(2020年5期)2020-05-19 09:41:34
    處暑
    村上春樹(shù)的17歲
    田原:閱讀是私密的堡壘
    女友·花園(2015年5期)2015-05-30 10:48:04
    新干線(xiàn)上的暗戀計(jì)劃
    篆刻·書(shū)畫(huà)
    村上春樹(shù)的17歲
    尋找左紳
    亚洲精品456在线播放app| av卡一久久| 美女cb高潮喷水在线观看| 亚洲成人一二三区av| 免费黄频网站在线观看国产| 久久女婷五月综合色啪小说 | 卡戴珊不雅视频在线播放| 韩国av在线不卡| 成人综合一区亚洲| 国产有黄有色有爽视频| 美女国产视频在线观看| 亚洲欧美清纯卡通| 一级爰片在线观看| 纵有疾风起免费观看全集完整版| 日韩av免费高清视频| 亚洲综合精品二区| 国产高清三级在线| 亚洲欧美一区二区三区黑人 | 日产精品乱码卡一卡2卡三| 在线观看三级黄色| av女优亚洲男人天堂| av线在线观看网站| 美女脱内裤让男人舔精品视频| 久久久久精品久久久久真实原创| 久久97久久精品| 日本黄色片子视频| 熟女av电影| 久久久久久久久久久免费av| .国产精品久久| 美女视频免费永久观看网站| 亚洲精品456在线播放app| 亚洲精品456在线播放app| 亚洲精品乱码久久久v下载方式| 亚洲精品成人av观看孕妇| 国产精品伦人一区二区| 日韩 亚洲 欧美在线| 久久久久国产精品人妻一区二区| 国产黄色视频一区二区在线观看| 人妻系列 视频| 成人亚洲精品av一区二区| 中文字幕免费在线视频6| 麻豆精品久久久久久蜜桃| 建设人人有责人人尽责人人享有的 | 国产乱人偷精品视频| www.av在线官网国产| 日日啪夜夜撸| 欧美精品人与动牲交sv欧美| www.色视频.com| 日本免费在线观看一区| 日韩欧美 国产精品| 免费高清在线观看视频在线观看| 男的添女的下面高潮视频| 日韩人妻高清精品专区| freevideosex欧美| 亚洲熟女精品中文字幕| av免费在线看不卡| 综合色av麻豆| 天美传媒精品一区二区| 69人妻影院| 一本久久精品| 日韩电影二区| 亚洲av中文av极速乱| 一级毛片电影观看| 国产精品伦人一区二区| 色网站视频免费| av免费在线看不卡| 乱系列少妇在线播放| 亚洲丝袜综合中文字幕| 日韩人妻高清精品专区| 最近中文字幕高清免费大全6| 久热久热在线精品观看| 色综合色国产| 97超视频在线观看视频| 国产人妻一区二区三区在| 联通29元200g的流量卡| 欧美精品一区二区大全| 久久久精品94久久精品| 国产成人a∨麻豆精品| 午夜福利视频1000在线观看| 男人舔奶头视频| 国产亚洲5aaaaa淫片| 国产精品av视频在线免费观看| 精品少妇久久久久久888优播| 97热精品久久久久久| 欧美3d第一页| 人妻少妇偷人精品九色| 久久久久网色| 在线观看人妻少妇| 国产精品99久久久久久久久| 日本与韩国留学比较| 丰满乱子伦码专区| 日韩 亚洲 欧美在线| 国产伦精品一区二区三区四那| 永久免费av网站大全| 亚洲自偷自拍三级| 国产亚洲精品久久久com| 蜜桃亚洲精品一区二区三区| 精品人妻熟女av久视频| 免费黄频网站在线观看国产| 国产老妇女一区| 日本wwww免费看| 嫩草影院新地址| 99热6这里只有精品| 国产高清三级在线| 人体艺术视频欧美日本| 亚洲av成人精品一二三区| 日日摸夜夜添夜夜爱| 91精品国产九色| 国产伦在线观看视频一区| 在线天堂最新版资源| 久久99热这里只有精品18| 国产高清有码在线观看视频| 国产精品一二三区在线看| 久久精品熟女亚洲av麻豆精品| 99视频精品全部免费 在线| 精品国产乱码久久久久久小说| 综合色丁香网| 国产成人aa在线观看| 国产淫语在线视频| 精品国产一区二区三区久久久樱花 | 97精品久久久久久久久久精品| 国产一区有黄有色的免费视频| 三级男女做爰猛烈吃奶摸视频| 久久精品国产自在天天线| 国产视频首页在线观看| 亚洲精品日本国产第一区| 成年女人在线观看亚洲视频 | 亚洲精品影视一区二区三区av| 午夜精品国产一区二区电影 | 亚洲国产精品专区欧美| 欧美日韩国产mv在线观看视频 | 日日啪夜夜爽| 搞女人的毛片| 亚洲精品乱码久久久久久按摩| 日本wwww免费看| 夫妻午夜视频| 亚洲高清免费不卡视频| 简卡轻食公司| 亚洲图色成人| 久热这里只有精品99| 精品国产一区二区三区久久久樱花 | 亚洲久久久久久中文字幕| 午夜免费男女啪啪视频观看| 伦理电影大哥的女人| 亚洲av欧美aⅴ国产| 99热网站在线观看| 青春草亚洲视频在线观看| 欧美成人a在线观看| 如何舔出高潮| 成人欧美大片| 欧美bdsm另类| 在线免费十八禁| 中文资源天堂在线| 日韩不卡一区二区三区视频在线| 久久久久精品性色| 少妇高潮的动态图| 夫妻午夜视频| 香蕉精品网在线| 欧美日韩一区二区视频在线观看视频在线 | 一本久久精品| 不卡视频在线观看欧美| 久久久国产一区二区| 蜜桃久久精品国产亚洲av| 欧美xxxx黑人xx丫x性爽| 国产av国产精品国产| 亚洲欧美精品专区久久| 国产精品人妻久久久影院| 男女边吃奶边做爰视频| 一区二区三区四区激情视频| 国产精品熟女久久久久浪| 免费黄网站久久成人精品| 波多野结衣巨乳人妻| 亚洲国产高清在线一区二区三| 亚洲精品视频女| 国产亚洲91精品色在线| 男男h啪啪无遮挡| 亚洲成人一二三区av| 精品视频人人做人人爽| 中国美白少妇内射xxxbb| 日韩人妻高清精品专区| 国模一区二区三区四区视频| 人妻一区二区av| 草草在线视频免费看| av女优亚洲男人天堂| 一个人看视频在线观看www免费| 久久久久久伊人网av| 丝袜脚勾引网站| 精品少妇黑人巨大在线播放| 黑人高潮一二区| 久久精品久久久久久久性| 99精国产麻豆久久婷婷| 校园人妻丝袜中文字幕| 久久精品熟女亚洲av麻豆精品| 亚洲天堂av无毛| 婷婷色综合www| 免费看不卡的av| 亚洲美女视频黄频| 国产精品国产三级国产专区5o| av天堂中文字幕网| 亚洲精品久久午夜乱码| 久热这里只有精品99| 精品国产乱码久久久久久小说| 久久久久网色| 精品国产一区二区三区久久久樱花 | 你懂的网址亚洲精品在线观看| 激情五月婷婷亚洲| 亚洲国产成人一精品久久久| 国产精品国产三级专区第一集| 成人无遮挡网站| 亚洲人成网站在线观看播放| 精品久久久久久电影网| 欧美丝袜亚洲另类| 麻豆久久精品国产亚洲av| 嫩草影院新地址| 欧美97在线视频| 日本色播在线视频| 国产淫片久久久久久久久| 久久鲁丝午夜福利片| 最近最新中文字幕免费大全7| 一区二区三区乱码不卡18| 又粗又硬又长又爽又黄的视频| 美女国产视频在线观看| 插阴视频在线观看视频| 99热这里只有是精品50| 美女脱内裤让男人舔精品视频| 国产一区二区亚洲精品在线观看| 国产精品久久久久久精品古装| .国产精品久久| 色哟哟·www| 99热全是精品| 伊人久久精品亚洲午夜| 亚洲国产av新网站| 国产亚洲av嫩草精品影院| 80岁老熟妇乱子伦牲交| 成人毛片a级毛片在线播放| 嫩草影院精品99| 成年av动漫网址| 边亲边吃奶的免费视频| 一级二级三级毛片免费看| 亚洲精品国产av蜜桃| 欧美xxⅹ黑人| 国产亚洲最大av| 欧美极品一区二区三区四区| 永久网站在线| 99九九线精品视频在线观看视频| 特大巨黑吊av在线直播| 插逼视频在线观看| 久久ye,这里只有精品| 特大巨黑吊av在线直播| 永久网站在线| 丰满乱子伦码专区| 蜜桃久久精品国产亚洲av| 1000部很黄的大片| av一本久久久久| 亚洲av不卡在线观看| 天堂中文最新版在线下载 | 久久精品久久精品一区二区三区| 中文在线观看免费www的网站| 在线观看av片永久免费下载| 成人漫画全彩无遮挡| 久热久热在线精品观看| 亚洲成人av在线免费| 国产视频首页在线观看| 久久久久久久久久久丰满| 国内精品美女久久久久久| 久久综合国产亚洲精品| av线在线观看网站| 成人亚洲欧美一区二区av| 中文精品一卡2卡3卡4更新| 亚洲精品第二区| 亚洲成色77777| 亚洲欧洲日产国产| 亚洲精品国产av蜜桃| 国产av码专区亚洲av| 免费av毛片视频| 一本一本综合久久| 看黄色毛片网站| 国产精品国产三级专区第一集| 国产成人免费无遮挡视频| 免费黄频网站在线观看国产| 日韩人妻高清精品专区| 久久韩国三级中文字幕| 精品久久久噜噜| 中文字幕av成人在线电影| 男男h啪啪无遮挡| 亚洲三级黄色毛片| 亚洲天堂av无毛| 日本一二三区视频观看| 91狼人影院| 少妇人妻久久综合中文| 老司机影院成人| 免费看不卡的av| 日韩中字成人| 色婷婷久久久亚洲欧美| 亚洲在线观看片| 你懂的网址亚洲精品在线观看| 爱豆传媒免费全集在线观看| 性色avwww在线观看| 国产精品伦人一区二区| 好男人视频免费观看在线| av在线播放精品| 久久精品久久久久久久性| 亚洲国产精品999| 菩萨蛮人人尽说江南好唐韦庄| 国产亚洲精品久久久com| 91狼人影院| 久久久久久伊人网av| 久久久色成人| 精品少妇黑人巨大在线播放| 我的女老师完整版在线观看| 在现免费观看毛片| 久久久精品免费免费高清| 精品熟女少妇av免费看| 国产综合精华液| 欧美一级a爱片免费观看看| 国产亚洲5aaaaa淫片| 国产亚洲精品久久久com| 国产精品99久久久久久久久| 最近最新中文字幕免费大全7| 女人被狂操c到高潮| 国产熟女欧美一区二区| 久久综合国产亚洲精品| 97在线人人人人妻| 91久久精品电影网| 久久99热这里只有精品18| 91在线精品国自产拍蜜月| 少妇人妻精品综合一区二区| 国产毛片a区久久久久| 少妇人妻久久综合中文| 韩国av在线不卡| 白带黄色成豆腐渣| xxx大片免费视频| 熟女人妻精品中文字幕| 亚洲,欧美,日韩| 国产国拍精品亚洲av在线观看| 超碰97精品在线观看| 全区人妻精品视频| 亚洲av一区综合| 中文在线观看免费www的网站| 国产精品一区二区性色av| 免费大片18禁| 新久久久久国产一级毛片| 寂寞人妻少妇视频99o| 高清日韩中文字幕在线| 99久久精品国产国产毛片| tube8黄色片| 亚洲av国产av综合av卡| 舔av片在线| 亚洲精品一区蜜桃| 精品熟女少妇av免费看| 亚洲精品乱久久久久久| 亚洲精品第二区| 免费看光身美女| 精品人妻视频免费看| 欧美zozozo另类| 老女人水多毛片| 亚洲真实伦在线观看| 日韩欧美精品免费久久| 国产成人一区二区在线| 观看美女的网站| 欧美日本视频| 一级毛片久久久久久久久女| 欧美+日韩+精品| 综合色丁香网| 少妇高潮的动态图| 国模一区二区三区四区视频| 欧美日韩一区二区视频在线观看视频在线 | 99久国产av精品国产电影| 亚洲国产高清在线一区二区三| 亚洲综合精品二区| 国产成人精品久久久久久| 99久国产av精品国产电影| 国产成人精品久久久久久| 国产精品熟女久久久久浪| 国内精品美女久久久久久| 老司机影院成人| 我的老师免费观看完整版| 国内精品宾馆在线| 22中文网久久字幕| 建设人人有责人人尽责人人享有的 | 九九在线视频观看精品| 插阴视频在线观看视频| av在线蜜桃| 好男人在线观看高清免费视频| 亚洲av日韩在线播放| 麻豆成人午夜福利视频| 天天躁夜夜躁狠狠久久av| 亚洲成人久久爱视频| 国产永久视频网站| 人妻系列 视频| 街头女战士在线观看网站| 又爽又黄无遮挡网站| 亚洲成人精品中文字幕电影| 国产91av在线免费观看| 99久久中文字幕三级久久日本| 国产色爽女视频免费观看| 五月玫瑰六月丁香| 可以在线观看毛片的网站| 精品久久久久久久久av| 免费看不卡的av| 久久人人爽av亚洲精品天堂 | 日本午夜av视频| 极品教师在线视频| 国产成人91sexporn| 久久久a久久爽久久v久久| 在线观看三级黄色| 看黄色毛片网站| 欧美性感艳星| 黄色配什么色好看| 亚洲综合精品二区| 午夜福利高清视频| 热re99久久精品国产66热6| 日本午夜av视频| 丝瓜视频免费看黄片| www.色视频.com| 99热这里只有是精品50| 激情五月婷婷亚洲| 免费看光身美女| av卡一久久| 国产精品久久久久久久电影| 下体分泌物呈黄色| 中文欧美无线码| 中文字幕亚洲精品专区| 免费av毛片视频| 国产成人精品一,二区| 有码 亚洲区| 日韩在线高清观看一区二区三区| 校园人妻丝袜中文字幕| 成人亚洲欧美一区二区av| 免费观看的影片在线观看| av在线蜜桃| 97精品久久久久久久久久精品| 自拍偷自拍亚洲精品老妇| 久久久久久久大尺度免费视频| 草草在线视频免费看| 免费黄频网站在线观看国产| 97在线视频观看| 伦精品一区二区三区| av国产精品久久久久影院| .国产精品久久| 日本欧美国产在线视频| 国产精品国产三级专区第一集| 美女视频免费永久观看网站| 美女国产视频在线观看| 秋霞伦理黄片| 嫩草影院新地址| 三级男女做爰猛烈吃奶摸视频| 亚洲av免费高清在线观看| 欧美精品人与动牲交sv欧美| 亚洲精品中文字幕在线视频 | 夫妻性生交免费视频一级片| 毛片一级片免费看久久久久| 国产探花极品一区二区| 国产精品爽爽va在线观看网站| 成人美女网站在线观看视频| 看非洲黑人一级黄片| 色视频www国产| 91久久精品电影网| 成年人午夜在线观看视频| 大陆偷拍与自拍| 免费观看无遮挡的男女| 十八禁网站网址无遮挡 | 精品少妇黑人巨大在线播放| 久久ye,这里只有精品| 日韩一区二区视频免费看| 777米奇影视久久| av在线老鸭窝| 高清日韩中文字幕在线| 久久精品国产a三级三级三级| 又大又黄又爽视频免费| 免费在线观看成人毛片| 插逼视频在线观看| 18禁在线无遮挡免费观看视频| 国产探花极品一区二区| 在线观看美女被高潮喷水网站| 91精品一卡2卡3卡4卡| 91久久精品国产一区二区三区| 国产探花在线观看一区二区| 国产男人的电影天堂91| 一区二区三区四区激情视频| 看免费成人av毛片| 毛片女人毛片| 免费看a级黄色片| 国产爽快片一区二区三区| 各种免费的搞黄视频| 18+在线观看网站| 国产av国产精品国产| 免费观看在线日韩| 在线免费观看不下载黄p国产| 亚洲精品成人久久久久久| 新久久久久国产一级毛片| 久久久久久国产a免费观看| 欧美人与善性xxx| av卡一久久| 永久免费av网站大全| 国产精品久久久久久精品古装| 日日撸夜夜添| 欧美老熟妇乱子伦牲交| 成人无遮挡网站| 91午夜精品亚洲一区二区三区| 蜜桃亚洲精品一区二区三区| 一本色道久久久久久精品综合| 18+在线观看网站| 少妇猛男粗大的猛烈进出视频 | 麻豆精品久久久久久蜜桃| 日本wwww免费看| 99久久精品一区二区三区| 69人妻影院| 全区人妻精品视频| 久久人人爽人人片av| 亚洲怡红院男人天堂| 美女被艹到高潮喷水动态| 在线观看美女被高潮喷水网站| 99精国产麻豆久久婷婷| 麻豆乱淫一区二区| 亚洲国产av新网站| 国产午夜精品一二区理论片| 美女内射精品一级片tv| 91午夜精品亚洲一区二区三区| 中文字幕免费在线视频6| 大片电影免费在线观看免费| 亚洲在久久综合| 日本欧美国产在线视频| 亚洲天堂av无毛| 亚洲在线观看片| 日韩成人av中文字幕在线观看| 亚洲电影在线观看av| 黄色视频在线播放观看不卡| 日韩中字成人| 男人添女人高潮全过程视频| 亚洲,一卡二卡三卡| 三级国产精品欧美在线观看| 晚上一个人看的免费电影| 五月玫瑰六月丁香| 亚洲欧美日韩东京热| 亚洲最大成人手机在线| 亚洲自拍偷在线| 国产亚洲av片在线观看秒播厂| av在线天堂中文字幕| 在线天堂最新版资源| 美女高潮的动态| av免费在线看不卡| 欧美人与善性xxx| 亚洲四区av| 亚洲精品成人久久久久久| 男人添女人高潮全过程视频| 日日摸夜夜添夜夜爱| 国产永久视频网站| 国产在线一区二区三区精| 噜噜噜噜噜久久久久久91| 人体艺术视频欧美日本| 国产69精品久久久久777片| 我的女老师完整版在线观看| 欧美+日韩+精品| 老司机影院成人| 精品国产乱码久久久久久小说| 午夜免费观看性视频| 少妇猛男粗大的猛烈进出视频 | 亚洲国产成人一精品久久久| 日本猛色少妇xxxxx猛交久久| 五月开心婷婷网| 熟女人妻精品中文字幕| 亚洲伊人久久精品综合| av在线亚洲专区| 国内精品宾馆在线| 日韩伦理黄色片| 赤兔流量卡办理| 好男人视频免费观看在线| 亚洲精品,欧美精品| 国产亚洲av片在线观看秒播厂| 亚洲av一区综合| 永久免费av网站大全| 九九爱精品视频在线观看| 欧美少妇被猛烈插入视频| 亚洲av二区三区四区| 三级国产精品片| 五月玫瑰六月丁香| 伦理电影大哥的女人| 秋霞在线观看毛片| av免费在线看不卡| 亚洲欧美日韩东京热| 国产成人aa在线观看| 日日撸夜夜添| 免费看光身美女| 国产成人aa在线观看| 国产成人免费观看mmmm| 不卡视频在线观看欧美| 熟女人妻精品中文字幕| 亚洲国产精品专区欧美| 欧美激情国产日韩精品一区| 简卡轻食公司| 日日撸夜夜添| 亚洲最大成人中文| 久久久久九九精品影院| 日韩一本色道免费dvd| 天堂俺去俺来也www色官网| 亚洲美女视频黄频| 国产成人免费无遮挡视频| 欧美日本视频| 欧美性感艳星| 欧美最新免费一区二区三区| 欧美成人a在线观看| 两个人的视频大全免费| 欧美丝袜亚洲另类| 成人毛片a级毛片在线播放| 蜜桃亚洲精品一区二区三区| 国产成年人精品一区二区| 欧美高清成人免费视频www| 亚洲国产精品成人综合色| 中文天堂在线官网| 欧美极品一区二区三区四区| 又粗又硬又长又爽又黄的视频| 亚洲,欧美,日韩| 99久久精品国产国产毛片| av播播在线观看一区|