• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    User Purchase Intention Prediction Based on Improved Deep Forest

    2024-01-20 13:01:52YifanZhangQianchengYuandLisiZhang

    Yifan Zhang ,Qiancheng Yu,2,★ and Lisi Zhang

    1School of Computer Science and Engineering,North Minzu University,Yinchuan,750030,China

    2Laboratory of Graphics and Images of the State Ethnic Affairs Commission,North Minzu University,Yinchuan,750030,China

    ABSTRACT Widely used deep neural networks currently face limitations in achieving optimal performance for purchase intention prediction due to constraints on data volume and hyperparameter selection.To address this issue,based on the deep forest algorithm and further integrating evolutionary ensemble learning methods,this paper proposes a novel Deep Adaptive Evolutionary Ensemble (DAEE) model.This model introduces model diversity into the cascade layer,allowing it to adaptively adjust its structure to accommodate complex and evolving purchasing behavior patterns.Moreover,this paper optimizes the methods of obtaining feature vectors,enhancement vectors,and prediction results within the deep forest algorithm to enhance the model’s predictive accuracy.Results demonstrate that the improved deep forest model not only possesses higher robustness but also shows an increase of 5.02%in AUC value compared to the baseline model.Furthermore,its training runtime speed is 6 times faster than that of deep models,and compared to other improved models,its accuracy has been enhanced by 0.9%.

    KEYWORDS Purchase prediction;deep forest;differential evolution algorithm;evolutionary ensemble learning;model selection

    1 Introduction

    In the era of the Internet,e-commerce has been widely adopted due to its convenience and efficiency[1].Against this backdrop,e-commerce platforms generate a wealth of user behavior data.Through the application of deep learning techniques[2]to analyze and predict these data,businesses can delve into the shopping intentions and demands of consumers,thereby offering more precise,personalized services and products.This can further increase customer satisfaction and business competitiveness.Among the numerous deep learning models,Deep Neural Networks (DNNs) have been extensively utilized due to their excellent predictive performance.However,their training process requires a large amount of hyperparameter tuning and lacks sufficient interpretability[3],undoubtedly increasing the complexity of business decision-making and understanding of user behavior patterns.

    In recent years,the Deep Forest(DF)[4]model,with its efficient computational performance and good model interpretability,has attracted the attention of researchers and businesses.Compared to Deep Neural Networks,Deep Forest exhibits significant advantages in processing different scales of data and model interpretability.However,despite the application of the Deep Forest model in scenarios such as user purchase intention prediction[5,6],there is still room for improvement in its predictive performance and model generalization ability.Therefore,this paper proposes an improved Deep Forest model to further enhance the accuracy of purchase intention prediction.

    The main contributions of this paper are as follows:

    · The Binary Differential Evolution algorithm is used for model selection.This provides a broader model selection space for Deep Forest,thereby accommodating complex data patterns.

    · A feature importance weighting strategy is adopted to operate on the original features in the cascading structure.This strategy exhibits efficient performance in guiding the training process of the next cascading forest,not only significantly improving the performance of the model but also ensuring its interpretability.

    · Meta-classifiers are introduced at each layer of the cascading forest.This strategy allows the model to further optimize the prediction results of the previous layer at each layer,thereby significantly enhancing the model’s predictive accuracy.

    · A deep ensemble strategy based on ensemble learning is implemented in the model to enhance its robustness and generalization capability.

    The remainder of this paper is organized as follows.In Section 2,we will discuss related work.Section 3 introduces the relevant theories needed for the article.In Section 4,we present the specific design of the model.In Section 5,we present the simulation results and analysis.Finally,the conclusions and prospects of this paper are presented.

    2 Related Work

    In the field of e-commerce,predicting user purchase intentions is a crucial area of research.With the rapid growth and complexity of data,traditional machine learning models,such as Logistic Regression (LR) [7],are often too simplistic,have weaker generalization capabilities,and limit improvements in predictive performance.Hybrid models,which combine various models or algorithms to produce a collective prediction result,have been applied.For instance,Tang et al.[8] combined Support Vector Machine (SVM) and Firefly Algorithm (FA),while Hu et al.[9] used SVM and LR together.These combinations effectively resolved the inadequacy of a single traditional model’s fitting capabilities.However,the combining process of these hybrid models is rather rudimentary and lacks flexibility,making further optimization challenging.Consequently,ensemble learning was introduced and has been widely applied in various prediction tasks.The main models include Random Forest (RF) [10] ensemble by bagging [11] method,and models like Extreme Gradient Boosting(XGBoost,XGB)[12],Light Gradient Boosting Machine(LightGBM,LGB)[13],Adaptive Boosting(AdaBoost,AdaB) [14],CatBoost (CatB) [15] and Gradient Boosting Decision Tree (GBDT) [16]that ensemble by boosting [17] method.These models integrate multiple different machine learning algorithms,overcoming the limitations of single algorithms.They offer high prediction accuracy,fast training speed,and low memory consumption,becoming mainstream solutions for data mining tasks[18].However,these ensemble learning methods still rely on human-defined model designs and feature selection,often requiring extensive experiments or expert knowledge to acquire the optimal model structure and parameter configuration.This limits their application in complex and largescale problems.To overcome these limitations,the concept of Evolutionary Ensemble Learning [19]was introduced.Evolutionary Ensemble Learning incorporates evolutionary algorithms like Genetic Algorithm(GA)[20]and Differential Evolution Algorithm(DE)[21],allowing for the automation and data-driven optimization of models.Wen et al.[22]proposed an ensemble learning framework based on a voting mechanism.This framework employs the voting mechanism for classification decisions and integrates the advantages of various constraint-handling techniques.To further enhance the performance of ensemble learning,a DE variant is used as the search engine,combined with four search strategies to generate new individuals,ensuring a balance between population diversity and convergence.

    Despite these models being part of shallow machine learning,they can enhance prediction accuracy but still struggle to perfectly express complex relationships within non-structured data[23].The DF model,a combination of ensemble learning and deep learning,possesses both the high accuracy of ensemble learning and the robust expressive ability of deep learning while avoiding the drawbacks of deep neural networks.Therefore,as a deep level ensemble model,DF has gradually gained attention and shown immense potential in purchase prediction tasks.In this regard,Ge et al.[6]proposed a user purchase behavior prediction method based on DF;Fu et al.[5]introduced a multilayer heterogeneous ensemble algorithm,constructing a user purchase behavior prediction algorithm framework based on the DF model.Although they validated the effectiveness of their methods and frameworks through experiments,there is still room for further optimization.

    In light of this,this paper integrates the advantages of evolutionary ensemble learning and deep forests to propose a Deep Adaptive Evolutionary Ensemble model for predicting user purchase intentions.A series of innovative research has been conducted in areas such as base model selection in deep forests,feature weight adjustment strategies,and model hierarchy optimization strategies,providing new solutions for purchase intention prediction in the e-commerce domain.

    3 Theoretical Foundation

    3.1 Deep Forest

    The Deep Forest model consists of two parts:Multi-Grained Scanning and Cascade Forest.Multi-Grained Scanning improves the ability of deep learning to represent data effectively by applying different sampling windows.Cascade Forest,on the other hand,is inspired by the layer-by-layer processing strategy of original features in neural networks,as shown in Fig.1.In this cascading structure,each forest attempts to further optimize data representation based on the previous forest.

    Figure 1:Cascade forest workflow

    In the Cascade Forest,each level is composed of several Random Forests(blue)and Extra Trees(ET)(black),with each forest consisting of multiple decision trees.In a regular Random Forest,each decision tree,during splitting,selects the feature with the optimal Gini value from a random subset of features to split.Extra Trees differ in that when deciding how to split a node,they randomly select a subset of features from all features and then randomly select a split point(instead of finding the optimal split point)until each leaf node contains only one category or no more than 10 samples.In this way,each decision tree outputs a result in the form of a class vector.Suppose it is a binary classification problem;it generates a two-dimensional class vector.The mean of the output class vector results of all decision trees is taken as the decision result of the Random Forest.The class vector output by each forest will serve as an enhanced feature vector and,after being concatenated with the original feature vector,will serve as the input for the next level.This process is repeated until the last level of the cascade,where the mean of the outputs of the last forest is calculated,and the category corresponding to the maximum value is taken as the final prediction result.

    To prevent overfitting in the deep structure,each forest’s enhanced vector is obtained through K-fold cross-validation.When the cascade extends to a new level,the performance of the entire cascade will be evaluated through the validation set.If there is no significant performance improvement,the training process will end.Therefore,the advantage of Deep Forest is that it can automatically adjust the model complexity to adapt to different scales of data sets,which is more adaptable compared to neural networks.

    3.2 Differential Evolution Algorithm

    The differential evolution algorithm is a population-based heuristic search algorithm that simulates the process of cooperation and competition among individuals within a population.Due to its clear algorithmic structure,few control variables,strong adaptability,and rapid convergence rate,it has been widely applied across various domains[24,25].The main algorithmic flow is as follows:

    (1)Initialization of the Population

    Based on the characteristics of the target task,the problem is abstracted into a manageable search space,with the problems to be solved converted into individuals.Each individual in the initialized population is a solution within the search range.In the solution space,Mindividuals are randomly and uniformly generated.Each individual is composed of an n-dimensional vector as the initial population,which is denoted as:

    The method of obtaining the jth dimension for the ith individual could be described as follows:

    (2)Evolutionary Operation

    1)Mutation Operation:

    For each parameter vector,three distinct individuals are randomly selected from the population to generate a mutant vector.The mutation equation is shown in Eq.(3).

    whereXr1(g),Xr2(g),andXr3(g)are three individuals selected from the population,selected randomly;r1,r2,andr3are integers andr1≠r2≠r3;Xr2(g)-Xr3(g)is the differential vector;grepresents the g-th generation;Fis the scale factor used to control the magnitude of mutation.

    2)Crossover Operation:

    The crossover operation involves recombining the target vectorXi,j(g)and mutant vectorVi,j(g)to produce a new trial vectorUi,j(g),thus retaining the information of the original individual and adding the information of the mutant individual.The specific equation is shown in Eq.(4).

    3)Selection Operation:

    The selection strategy is based on the assessment of the fitness of the trial vector and the original individual to select the strategy for the next generation.Its purpose is to select the optimal individual in the population so that the population gradually converges to the global optimal solution.The specific selection method is as follows:

    wherefrepresents the fitness function.A greedy selection strategy is adopted betweenUi(g)andXi(g),with the superior individual serving as the new individual.

    4 User Purchase Intention Prediction Method Design

    4.1 Model Selection of Binary Differential Evolution Algorithm

    The traditional Cascade Forest model mainly uses RF and ET as base models.However,both of these models use the same ensemble method at their core,which may not fully capture all the effective information in the prediction task.The goal of the Cascade Forest is to improve final performance by integrating predictions from multiple models.However,models based on a single structure may limit the complementarity between models,thereby affecting the model’s expressive power and generalization performance.Although some research has enhanced model diversity by introducing heterogeneous models[26,27],this strategy also has its undeniable limitations.For example,without strict model selection and optimization,it might introduce models that do not perform well on specific tasks or datasets,which could weaken the overall performance of the model.To overcome these problems,this paper proposes a method for multi-model optimized cascade forest structure,using a binary differential evolution algorithm for model selection,and introduces heterogeneous models into the cascade layer.

    Firstly,a population is initialized.If there areNclassifiers in the target task,then a 1*Nbinary sequence is initialized,where the value of each element is either 0 or 1.The coding method is shown in Fig.2.

    Figure 2:Binary differential evolution algorithm encoding method

    Based on the number of classifiers,an appropriate population sizeNPis set,and the above binary encoding is repeatedNPtimes to obtain a set containingNPbinary sequences.

    The fitness value for this classifier is defined as the mean of the Accuracy and Area Under the Curve(AUC)obtained after the Decision Tree classifier performs 5-fold cross-validation on the input data.Its definition is as follows:

    whereXrepresents the subset of classifiers represented by the binary vector;Accuracy5-fold(X) andAUC5-fold(X)refer to the Accuracy and AUC values obtained by performing 5-fold cross-validation of the classifier on the training data.

    After the mutation operation is performed,Eq.(7)is used to correct the mutation vector to ensure that each element is between 0 and 1.This guarantees that each element of the individual after mutation is a binary bit,ensuring the correctness of the algorithm.

    During multiple rounds of iteration,the algorithm will perform evolutionary operations(including mutation,and crossover operations) on the individuals of each round to generate new binary sequences(new individuals).Then,based on the scoring results of the fitness function,comparisons and selections are made between the newly generated individuals and the original individuals.Finally,when the binary differential evolution algorithm reaches the preset maximum number of iterations,the optimal model combination is selected from the predefined classifier set and passed to the cascade layer for use.

    4.2 Improved Cascade Forest Model

    For the deep forest model,it enhances the model’s prediction performance by performing layered feature extraction and combination through multi-level cascading.However,as the number of cascading layers increases,the problem of“feature degradation”starts to emerge.This is because the weight of the original features in the feature vector may gradually decrease [28],and even in subsequent cascade layers,they may be completely replaced by the feature vectors output from the first few layers.Ignoring the different contributions of each feature to the classification result and simply concatenating the original features with the output feature vectors of the cascade layer may“dilute”the information of important features with relatively less important features,which undoubtedly has a negative impact on the model performance.

    In binary classification problems like purchase intention prediction,each cascade layer’s forest will produce two category probability vectors(purchase and non-purchase),the sum of the probabilities of these two vectors is 1.These probability outputs will serve as inputs for the next level,leading to an increase in data dimensions[29,30].In addition,since the probability features newly added in each cascade layer are highly correlated with the existing features,the amount of information they add is relatively limited,leading to redundancy in the feature vector,and also increasing the computational cost of the model.

    Therefore,in response to the above problems,this paper has optimized the original feature vector and the output of prediction probability in a series of ways.As shown in Fig.3.

    Figure 3:Improved cascade forest structure

    In the processing of feature vectors,a feature weighting strategy is adopted to enhance the model’s prediction performance and interpretability.First,the importance of each feature is calculated in each ensemble-based classifier,and these importance values reflect the degree of contribution of each feature to the prediction result.Then,we calculate the average feature importance across all models,which serves as the feature weight.The corresponding equation is as follows:

    wherenis the number of base classifiers,wiis the weight of featurefi,and Imp(M,f)is the importance score assigned by the modelMto featuref,this process is implemented by the feature_importances_function in the sklearn library.

    Each feature fiwill be adjusted according to its corresponding weightwi,as shown in Eq.(9).

    In each layer of the cascading structure,the importance of features will be reevaluated and these features will be adjusted according to this.This process allows the feature weight calculation of subsequent layers to be based on the weighted features of the previous layer,thus achieving layerby-layer enhancement of important features.This cascaded weighting plays an important role in the cascade forest model,enhancing the model’s learning ability for important features,and allowing the model to more accurately use the information contained in these key features for prediction.In addition,reducing the model’s sensitivity to the evaluation errors of individual base classifiers significantly improves the robustness of the model.

    In the aspect of processing class probability vectors,this paper only uses the positive class(label 1)probability as the enhanced feature,different from the traditional approach(i.e.,considering both positive and negative class probabilities).For each cascade layer,a set of base classifiers are trained independently and generate their positive class probability predictions.Then,a meta-classifier is used to learn these positive class probabilities,understand the correlation between the prediction results of each base classifier,and output an integrated prediction probability.This prediction probability represents the cascade model’s prediction at that level and directly affects the output of the current layer.In addition,this prediction result is also introduced as a new enhanced feature into the next layer,forming a self-learning and self-optimization mechanism.In the training process,if there is no significant improvement in model performance in three consecutive cascade layers,the cascade forest process is terminated.Finally,the model’s prediction result is obtained by collecting the prediction probabilities of each meta-classifier at each level and taking their average.

    The improved cascade forest introduces the feature importance assessment of the base learner and dynamically adjusts the weights of the features at each level,making the model more focused on those features with higher discriminative power and reducing the noise impact of irrelevant features.At the same time,by training a group of base classifiers to output positive class probability predictions,and using a meta-classifier to learn and integrate prediction probabilities,effectively reducing feature dimensionality while incorporating information from all base classifiers.These strategies effectively integrate the predictive power of multiple algorithms,capturing the complexity of data at different levels,thereby improving the accuracy and robustness of model prediction.

    4.3 Overall Structure

    Following the above enhancements to the cascade structure,a Deep Adaptive Evolutionary Ensemble model was constructed for user purchase intention prediction in this paper.The architecture of this model is illustrated in Fig.4.

    Figure 4:Framework of the deep adaptive evolutionary ensemble learning model

    On a given dataset,an evolutionary ensemble learning strategy is employed in the model selection layer.This strategy optimizes the objective function to select a set of classifiers,and the results of these classifiers are then fed back into the cascade forest.In the cascade forest,if the initial input feature dimension is 400 and each cascade layer consists of four base classifiers(as determined by the model selection layer),the prediction probabilities from each base classifier are first concatenated together to form a matrix.This matrix serves as the input to the meta-classifier.The meta-classifier learns from this input data and predicts a target variable,which is then concatenated with the weighted feature vector.Therefore,at the end of the first layer,we get a feature vector of 401 dimensions.This vector serves as the input to the next layer.This process continues until the stopping condition for layer expansion is met.

    During the testing phase,the cascade structure is used to make predictions on the test instances,yielding the final layer’s prediction results.The final prediction is the average of the prediction probabilities obtained by the meta-classifier at each layer.If this average value is greater than or equal to 0.5,the instance is predicted to be a positive case;otherwise,it is predicted as negative.

    5 Experimental Analysis

    5.1 Data Description

    The experimental data used in this paper is from real online transaction data of JD.com,comprising 4 categories and a total of 6 datasets.For detailed information,please refer to Table 1.

    Table 1: Data set distribution

    JD Mall provided data from February 01,2016,to April 15,2016,including behavioral data of 105,321 users,totaling 50,601,736 records;at the same time,it also includes 558,552 review data of 24,187 products.

    5.2 Evaluation Metrics

    The experiment in this paper uses Accuracy,Precision,Recall,F1-score,and AUC as the performance evaluation indicators of the model.The calculation equations are shown as follows(Eqs.(10)to(14)):

    5.3 Experimental Environment and Parameter Settings

    Table 2 lists the software and hardware environments,model methods,etc.,used in the experiment.

    Table 2: Experimental environment configuration

    Table 3 provides detailed records of the parameter configuration information for all models used in this paper’s experiments.

    Table 3: Model parameter settings

    5.4 Experimental Results

    5.4.1ModelSelectionResults

    To implement the multi-model ensemble framework,this paper selects seven traditional ensemble models: RandomForest,XGBoost,CatBoost,LightGBM,AdaBoost,GBDT,and ET.The most suitable model combination for the purchasing intention dataset is determined by using the binary differential evolution algorithm for model selection.Upon reaching the preset maximum number of iterations,the optimal model combination{XGB,ET,RF}is obtained.

    As shown in Fig.5,in the evaluation of fitness values,the LightGBM model performs the best with a fitness value of 0.8458.This is closely followed by the XGBoost and RandomForest models,with fitness values of 0.8387 and 0.8332,respectively.These results demonstrate that the LightGBM model performs best in terms of adaptability and has a high classification accuracy.However,it is worth noting that despite LightGBM having the highest fitness value,it appears relatively infrequently in the best combination during the iterative process.This could be because the parameter setting and model architecture of LightGBM did not reach the optimum during the search process of the binary differential evolution algorithm.

    Figure 5:Fitness values and occurrences of each classifier during the model selection process

    5.4.2ComparisonExperimentwithBaseModels

    Table 4 thoroughly compares the performance of the DAEE model with the original Deep Forest model and seven traditional single ensemble models on five evaluation metrics.The experimental results indicate that,compared with these eight models,the DAEE model,as an architecture based on heterogeneous models and multi-level ensembles,has achieved significant performance improvements on all metrics except the precision indicator.Notably,the DAEE model scored 71.97% on the AUC metric,demonstrating its significant superiority in classification tasks compared to other models.

    Table 4: Comparison of performance prediction in feature selection experiments

    5.4.3ComparisonExperimentwithDeepModels

    In this experiment,several versions of deep neural networks are included:DNN_50,DNN_100,DNN_150,and DNN_200.The primary difference between these versions lies in the number of epochs for training,which are 50,100,150,and 200,respectively.

    From the experimental results in Table 5,it can be observed that as the number of training epochs for the DNN increases,the model’s performance on the test set shows an improving trend in the initial stages.However,when the number of epochs exceeds 150,the performance of the model declines.This could be due to overfitting during the training process,which adversely impacts the performance on the test set.Moreover,as the number of epochs increases,the model’s training time also correspondingly increases.

    Table 5: Comparison of prediction performance with deep learning models

    DAEE outperforms DNN_50,DNN_100,and DNN_200 on the AUC,F1,recall,and precision metrics.Compared to DNN_150,although the DAEE model scores slightly lower on all metrics,the runtime of DNN_150 is six times that of the DAEE model.This makes the DAEE model exhibit higher feasibility in real-time or large-scale application scenarios.

    5.4.4ComparisonExperimentofDifferentMeta-Classifiers

    To validate the stability of the DAEE model in adaptively adjusting the combination of base learners,selecting meta-classifiers,and optimizing the deep forest method,the best-selected learners were compared using them as meta-classifiers in experimental tests.In Table 6,DAE_XGB,DAE_RF,and DAE_ET respectively represent scenarios where the selected learners in the selection layer were used as meta-classifiers in sequence.The results show that the difference in AUC values between the best-performing and worst-performing meta-classifiers is 0.0107.This indicates that the proposed model’s overall performance remains excellent regardless of which meta-classifier is used.It further confirms the high stability and robustness of the DAEE model when handling purchase intention prediction data.

    Fig.6 shows the performance comparison of models with three different meta-classifiers and traditional models on four indicators: AUC,F1,recall,and Accuracy.Regardless of which optimal learner is chosen as the meta-classifier,the performance of the proposed model on all indicators is significantly better than that of the single ensemble model,fully verifying the effectiveness of our proposed Deep Evolutionary Ensemble Learning model.

    Table 6:Comparison of performance of deep adaptive ensemble model under different meta-classifiers

    Figure 6:Performance comparison between different meta-classifiers and other models

    RF and ET models,which are ensembled using Bagging,do not perform as well as models ensembled using Boosting.This might be due to the fact that the parallel sampling and resampling strategy of Bagging methods cannot effectively deal with complex and unbalanced purchase prediction problems.In contrast,models ensembled using the Boosting method can adjust sample weights based on training results [31],assigning higher weights to samples with incorrect predictions,hence performing relatively better on the indicators.It is worth noting that when RF and ET,together with the XGBoost model,which is ensembled using Boosting,are inputted into the Deep Forest as the optimal model combination,good results are achieved on all indicators.This result indicates that through model diversity,layered training,and self-enhancement mechanisms,the prediction performance of the model can be significantly improved,further demonstrating the excellent performance of the proposed model in prediction tasks.

    5.4.5AblationExperiment

    To verify the effectiveness of the DAEE model in adaptively introducing heterogeneous models,handling meta-classifier prediction probabilities,and adjusting feature weight strategies,an ablation study was conducted in this paper.The experimental configurations are as follows:

    (1)Eliminating the feature importance weighting operation,denoted as w/o FI.

    (2) Not using the meta-classifier prediction probability,but directly using the average of the classifier prediction values as output,labeled as w/o ML.

    (3)Simultaneously ignoring feature importance and meta-classifier prediction probability,marked as w/o FI&ML.

    (4)Excluding the model selection layer and only using the RF and DT models to predict within the improved cascading layer,denoted as w/o MS.

    For detailed experimental results,refer to Table 7.

    Table 7: Ablation study results

    From Table 7,it is evident that when not using the meta-classifier prediction probability,the model’s prediction performance(measured by the AUC value)decreased by 3.55%.When relying solely on the improved cascading layer for prediction,with RF and DT acting as intra-layer classifiers,the AUC dropped by 5.04%.In scenarios without feature weighting,the model’s predictive performance declined by 6.42%,highlighting the crucial role of feature weight in model prediction.When both feature importance and meta-classifier prediction probability are disregarded,the prediction performance exhibits the most significant decline.However,thanks to the heterogeneous models integrated within the cascading layer,its performance still surpasses that of the original DF model.

    5.4.6ComparisonExperimentwithOtherImprovedMethods

    There is a wealth of research on improving the cascade forest part of the Deep Forest model.References [29] and [30] optimized the enhancement vectors: the former designed a strong cascade forest model that selects class vectors generated at each level to achieve effective representation,while the latter used the mean value of the same class classification probability results output by similar forest models as input to achieve dimensionality reduction.Reference[28]introduced the AdaBoost algorithm to enhance the self-adjusting ability of the Deep Forest model and recombined it with the mean of the class distribution vectors output by each cascade layer after updating the weights of the features according to the classification error rate at each level,achieving simultaneous optimization of features and enhancement vectors.This paper introduces these three algorithms for comparison with the DF and DAEE models.

    As shown in parts(a)and(b)of Fig.7,we compare the DAEE model with the benchmark model DF21 and the models proposed in references[28,29],and[30]on the AUC and Accuracy indicators.On the AUC indicator,the performance of the DAEE model surpassed the benchmark model DF and the models in references[29]and[30],but was slightly lower than the model in reference[28].This result indicates that the strategy of simultaneously adjusting original features and enhancement vectors can effectively improve the performance of the model.Similarly,the DAEE model also performs well on the accuracy indicator.

    Figure 7:Comparison experiment with other improved methods.(a)AUC metrics;(b)accuracy metrics

    6 Conclusion and Prospect

    To address the limitations of deep neural networks in terms of data volume and hyperparameters,and to achieve adaptive optimization in the model structure,this paper introduces a Deep Adaptive Evolutionary Ensemble model applied to user purchase intention prediction.This model consists of two core components:the model selection layer and the cascading layer.In the model selection layer,a binary differential evolution algorithm is utilized to adaptively select models,and the chosen highperforming models are then integrated into the cascading forest layer-by-layer.In the cascading layer,strategies based on positive class probability enhanced features,meta-classifier prediction probability enhancement,and feature weight adjustment are applied to further optimize the original cascading structure.Experimental results show that compared to single ensemble models,this model has higher classification performance,improving the AUC value by 5.02%.In terms of training speed,this model is six times faster than deep models.Furthermore,its prediction accuracy increased by 0.9%compared to other optimization schemes.In summary,the model proposed in this paper effectively integrates the advantages of various machine learning techniques,and through ensemble-based deep integration and various optimization strategies,it significantly enhances learning efficiency and predictive performance when dealing with complex purchase prediction problems.In future work,various intelligent algorithms will be explored to validate the superiority of the proposed model.

    Acknowledgement:The author thanks the School of Computer Science and Engineering of North Minzu University for providing equipment support.

    Funding Statement:This paper is financially supported by Ningxia Key R&D Program (Key)Project (2023BDE02001);Ningxia Key R&D Program (Talent Introduction Special) Project(2022YCZX0013);North Minzu University 2022 School-Level Research Platform “Digital Agriculture Empowering Ningxia Rural Revitalization Innovation Team”,Project Number:2022PT_S10;Yinchuan City School-Enterprise Joint Innovation Project (2022XQZD009);“Innovation Team for Imaging and Intelligent Information Processing”of the National Ethnic Affairs Commission.

    Author Contributions:Study conception and design:Y.Z.,Q.Y.;data processing:Y.Z.and L.Z.;draft manuscript preparation:Y.Z.,L.Z.;correction of errors and writing instructions:Q.Y.

    Availability of Data and Materials:The experimental data used in this study is actual online transaction data from JD.com,which has been anonymized for the purpose of protecting user privacy.Please visit the following link to access the data:https://www.datafountain.cn/competitions/247/datasets.

    Conflicts of Interest:The authors declare they have no conflicts of interest to report regarding the present study.

    露出奶头的视频| 成人欧美大片| 亚洲专区中文字幕在线| 黄片小视频在线播放| 男女之事视频高清在线观看| 免费看日本二区| 国产激情欧美一区二区| 99久久久亚洲精品蜜臀av| 男女下面进入的视频免费午夜| 国产精品女同一区二区软件 | 国产亚洲精品av在线| 亚洲精品一卡2卡三卡4卡5卡| 97人妻精品一区二区三区麻豆| 国产精品九九99| 一本久久中文字幕| 国产成人aa在线观看| 99久久国产精品久久久| 最新在线观看一区二区三区| x7x7x7水蜜桃| 首页视频小说图片口味搜索| 制服人妻中文乱码| 岛国在线观看网站| 97碰自拍视频| 一进一出好大好爽视频| 国产亚洲av嫩草精品影院| 国产成人啪精品午夜网站| 国产亚洲精品av在线| 日韩免费av在线播放| 国产高清激情床上av| 日日干狠狠操夜夜爽| 日本黄色片子视频| 亚洲va日本ⅴa欧美va伊人久久| 国产男靠女视频免费网站| 色综合站精品国产| 最新在线观看一区二区三区| 天天一区二区日本电影三级| 亚洲自拍偷在线| 亚洲成人中文字幕在线播放| 国产欧美日韩一区二区三| 国产一区二区激情短视频| 露出奶头的视频| 老鸭窝网址在线观看| 无人区码免费观看不卡| 亚洲男人的天堂狠狠| 怎么达到女性高潮| 欧美日韩瑟瑟在线播放| 一二三四在线观看免费中文在| 国产精品香港三级国产av潘金莲| 91久久精品国产一区二区成人 | svipshipincom国产片| 国产在线精品亚洲第一网站| 国产一区二区三区视频了| 精品国产美女av久久久久小说| 搡老妇女老女人老熟妇| 国产美女午夜福利| 国产成人啪精品午夜网站| 99久久成人亚洲精品观看| 天天一区二区日本电影三级| 99久久精品热视频| 日韩av在线大香蕉| 色综合欧美亚洲国产小说| 国产免费男女视频| 国产蜜桃级精品一区二区三区| 亚洲国产欧美一区二区综合| 九色国产91popny在线| 九九在线视频观看精品| 久久精品国产亚洲av香蕉五月| 啪啪无遮挡十八禁网站| 99热这里只有精品一区 | 琪琪午夜伦伦电影理论片6080| 亚洲人与动物交配视频| 一个人看视频在线观看www免费 | 亚洲五月天丁香| 精品99又大又爽又粗少妇毛片 | 国产精品1区2区在线观看.| 亚洲欧美日韩无卡精品| 日韩av在线大香蕉| 老司机午夜福利在线观看视频| 欧洲精品卡2卡3卡4卡5卡区| 亚洲男人的天堂狠狠| 一级毛片女人18水好多| 亚洲午夜理论影院| h日本视频在线播放| 久久精品人妻少妇| 黄色女人牲交| 午夜福利免费观看在线| 亚洲国产欧美一区二区综合| 国内精品久久久久久久电影| 丰满人妻熟妇乱又伦精品不卡| 人妻丰满熟妇av一区二区三区| 亚洲国产日韩欧美精品在线观看 | 亚洲五月婷婷丁香| 美女被艹到高潮喷水动态| 久久久久久九九精品二区国产| 成年女人永久免费观看视频| 欧美成狂野欧美在线观看| 老司机福利观看| 国产人伦9x9x在线观看| 精品久久久久久久久久久久久| 天堂av国产一区二区熟女人妻| 日韩国内少妇激情av| 麻豆成人av在线观看| 国产三级中文精品| 一区福利在线观看| 国产精品九九99| 亚洲精华国产精华精| 国产成人一区二区三区免费视频网站| 免费看日本二区| 免费在线观看日本一区| 黄色日韩在线| 99re在线观看精品视频| 91久久精品国产一区二区成人 | 精品日产1卡2卡| 国产午夜精品久久久久久| 日韩欧美一区二区三区在线观看| 中文字幕人妻丝袜一区二区| 91av网站免费观看| 久久午夜亚洲精品久久| 国产成人啪精品午夜网站| 亚洲无线观看免费| 免费在线观看影片大全网站| 国产伦精品一区二区三区视频9 | 国产精品亚洲美女久久久| 国产精品99久久久久久久久| 成人特级黄色片久久久久久久| av天堂在线播放| 91麻豆精品激情在线观看国产| 国产精品自产拍在线观看55亚洲| 这个男人来自地球电影免费观看| 精品无人区乱码1区二区| 中文字幕最新亚洲高清| 国产熟女xx| 午夜福利在线观看免费完整高清在 | 久久国产乱子伦精品免费另类| 国产欧美日韩精品亚洲av| 亚洲性夜色夜夜综合| 禁无遮挡网站| 久久久久国产一级毛片高清牌| 丁香六月欧美| 国产伦在线观看视频一区| 国产爱豆传媒在线观看| 国产伦精品一区二区三区视频9 | 国产主播在线观看一区二区| 露出奶头的视频| 香蕉国产在线看| 亚洲男人的天堂狠狠| 成人永久免费在线观看视频| 久久国产乱子伦精品免费另类| 亚洲七黄色美女视频| 两人在一起打扑克的视频| 亚洲av片天天在线观看| 国内精品美女久久久久久| 国产伦精品一区二区三区视频9 | 国产三级在线视频| 一区福利在线观看| 成年女人永久免费观看视频| 欧美黑人巨大hd| 午夜激情福利司机影院| 亚洲欧洲精品一区二区精品久久久| 国内精品一区二区在线观看| 97碰自拍视频| 亚洲国产欧洲综合997久久,| 欧美在线一区亚洲| 免费在线观看亚洲国产| 最近最新中文字幕大全电影3| 亚洲无线观看免费| 又粗又爽又猛毛片免费看| 亚洲熟妇中文字幕五十中出| 成人无遮挡网站| 精品99又大又爽又粗少妇毛片 | 免费在线观看亚洲国产| 看片在线看免费视频| 日本黄色视频三级网站网址| 日本 av在线| a级毛片a级免费在线| 丝袜人妻中文字幕| 少妇的丰满在线观看| 免费在线观看日本一区| 女同久久另类99精品国产91| 亚洲国产精品999在线| 亚洲av免费在线观看| 天天躁日日操中文字幕| 日本撒尿小便嘘嘘汇集6| 精品久久久久久久久久免费视频| 日韩国内少妇激情av| 亚洲乱码一区二区免费版| 精品久久久久久久毛片微露脸| 精品无人区乱码1区二区| 天堂av国产一区二区熟女人妻| 小说图片视频综合网站| 午夜影院日韩av| 中文字幕熟女人妻在线| 亚洲人成伊人成综合网2020| 中文字幕高清在线视频| 国内少妇人妻偷人精品xxx网站 | 淫妇啪啪啪对白视频| 欧美精品啪啪一区二区三区| 丁香欧美五月| 90打野战视频偷拍视频| 亚洲成人久久爱视频| 中文字幕久久专区| 日本黄色视频三级网站网址| 在线免费观看的www视频| 狠狠狠狠99中文字幕| 免费看日本二区| 又黄又粗又硬又大视频| 午夜福利18| 日本免费a在线| 国产亚洲欧美98| 欧美日韩国产亚洲二区| av中文乱码字幕在线| or卡值多少钱| 99热6这里只有精品| 亚洲中文日韩欧美视频| 大型黄色视频在线免费观看| 国产亚洲欧美在线一区二区| 亚洲国产色片| 午夜久久久久精精品| 亚洲中文字幕一区二区三区有码在线看 | 日日干狠狠操夜夜爽| 亚洲第一欧美日韩一区二区三区| 好看av亚洲va欧美ⅴa在| 日本与韩国留学比较| 欧美性猛交╳xxx乱大交人| 无人区码免费观看不卡| 欧美大码av| 又黄又爽又免费观看的视频| 一a级毛片在线观看| 欧美激情在线99| 欧美中文综合在线视频| 黄色成人免费大全| 男女下面进入的视频免费午夜| 91av网站免费观看| 日本黄色片子视频| 两性午夜刺激爽爽歪歪视频在线观看| 99re在线观看精品视频| 叶爱在线成人免费视频播放| 免费在线观看成人毛片| 国产亚洲精品av在线| 国产一级毛片七仙女欲春2| 国产精品久久久av美女十八| 日韩欧美免费精品| www国产在线视频色| 亚洲国产精品合色在线| 色在线成人网| 91在线精品国自产拍蜜月 | 伦理电影免费视频| 国产精品久久久久久久电影 | 久久中文字幕人妻熟女| 国产97色在线日韩免费| 国产精品99久久久久久久久| 97人妻精品一区二区三区麻豆| 亚洲欧美日韩卡通动漫| 国产又黄又爽又无遮挡在线| 久久精品人妻少妇| 最近视频中文字幕2019在线8| 久久久久性生活片| 久久精品亚洲精品国产色婷小说| 美女大奶头视频| 国产精品久久久久久人妻精品电影| 成人三级做爰电影| 在线观看66精品国产| 国产91精品成人一区二区三区| tocl精华| 成人一区二区视频在线观看| 97碰自拍视频| 欧美日韩乱码在线| 日韩欧美在线二视频| 日本成人三级电影网站| 天堂动漫精品| 18禁黄网站禁片午夜丰满| 国产精品99久久久久久久久| 在线观看一区二区三区| 麻豆av在线久日| 天堂网av新在线| 伊人久久大香线蕉亚洲五| av在线天堂中文字幕| 老司机福利观看| 三级男女做爰猛烈吃奶摸视频| 国产主播在线观看一区二区| 午夜免费激情av| 日日夜夜操网爽| 2021天堂中文幕一二区在线观| av视频在线观看入口| 国产高清videossex| 欧美日韩黄片免| 亚洲18禁久久av| 非洲黑人性xxxx精品又粗又长| 亚洲专区字幕在线| 久久精品夜夜夜夜夜久久蜜豆| 男插女下体视频免费在线播放| 日日摸夜夜添夜夜添小说| 婷婷精品国产亚洲av在线| 男女之事视频高清在线观看| 亚洲国产精品久久男人天堂| 国产一级毛片七仙女欲春2| 亚洲无线在线观看| 午夜福利成人在线免费观看| 欧美日韩综合久久久久久 | 午夜福利在线在线| 极品教师在线免费播放| 国产成人啪精品午夜网站| 亚洲一区二区三区色噜噜| 黄色视频,在线免费观看| 亚洲国产精品久久男人天堂| 不卡一级毛片| 在线观看午夜福利视频| 亚洲午夜精品一区,二区,三区| 久久久精品欧美日韩精品| 啦啦啦观看免费观看视频高清| 久久99热这里只有精品18| 精品久久蜜臀av无| 精品久久久久久,| 亚洲人成伊人成综合网2020| 一进一出抽搐gif免费好疼| 青草久久国产| 夜夜看夜夜爽夜夜摸| а√天堂www在线а√下载| 黑人操中国人逼视频| 九色成人免费人妻av| 国产aⅴ精品一区二区三区波| 中文字幕人成人乱码亚洲影| 国产v大片淫在线免费观看| 亚洲精品一区av在线观看| 欧美日韩福利视频一区二区| 黄色 视频免费看| 嫁个100分男人电影在线观看| 黄色日韩在线| 在线观看午夜福利视频| 午夜两性在线视频| 嫁个100分男人电影在线观看| 免费av不卡在线播放| cao死你这个sao货| h日本视频在线播放| 国产成人精品无人区| 两性午夜刺激爽爽歪歪视频在线观看| 亚洲av第一区精品v没综合| 亚洲国产日韩欧美精品在线观看 | 99热6这里只有精品| 老熟妇仑乱视频hdxx| 一本一本综合久久| 成人av一区二区三区在线看| 别揉我奶头~嗯~啊~动态视频| 亚洲熟女毛片儿| 国产精品 国内视频| 亚洲最大成人中文| 色播亚洲综合网| 亚洲va日本ⅴa欧美va伊人久久| 偷拍熟女少妇极品色| 色综合婷婷激情| 亚洲精品国产精品久久久不卡| 脱女人内裤的视频| 成年女人永久免费观看视频| 亚洲中文日韩欧美视频| 天天添夜夜摸| 一个人看视频在线观看www免费 | 亚洲国产中文字幕在线视频| 久久精品aⅴ一区二区三区四区| 人妻久久中文字幕网| 丁香欧美五月| 亚洲国产中文字幕在线视频| 色综合亚洲欧美另类图片| 国产激情久久老熟女| av欧美777| 国产激情久久老熟女| 午夜福利欧美成人| 日韩欧美在线二视频| 热99在线观看视频| 亚洲国产精品久久男人天堂| 国产私拍福利视频在线观看| 男女那种视频在线观看| 99精品欧美一区二区三区四区| 99热这里只有是精品50| 亚洲精品美女久久久久99蜜臀| 亚洲欧美日韩高清在线视频| 99在线人妻在线中文字幕| 精品国产乱子伦一区二区三区| 在线免费观看的www视频| 国产高清三级在线| 美女大奶头视频| 久久久久国产一级毛片高清牌| 丁香欧美五月| 日本三级黄在线观看| 久久亚洲精品不卡| 制服人妻中文乱码| 长腿黑丝高跟| 波多野结衣巨乳人妻| 欧美日韩中文字幕国产精品一区二区三区| 国产亚洲精品av在线| 国内毛片毛片毛片毛片毛片| 亚洲国产精品合色在线| 18美女黄网站色大片免费观看| 黑人巨大精品欧美一区二区mp4| 99riav亚洲国产免费| 精品国产亚洲在线| 国产美女午夜福利| 午夜福利18| 亚洲欧美精品综合久久99| 在线观看一区二区三区| 制服人妻中文乱码| 国产黄a三级三级三级人| 小蜜桃在线观看免费完整版高清| 舔av片在线| 欧美日韩国产亚洲二区| 日韩av在线大香蕉| 亚洲中文字幕日韩| 可以在线观看的亚洲视频| 国产精品久久久久久亚洲av鲁大| 欧美黄色淫秽网站| 在线观看一区二区三区| 丰满人妻熟妇乱又伦精品不卡| 色吧在线观看| 丁香六月欧美| 在线观看美女被高潮喷水网站 | 91麻豆av在线| 国产精品久久久人人做人人爽| 最新美女视频免费是黄的| 一二三四社区在线视频社区8| 国产伦精品一区二区三区四那| a级毛片a级免费在线| 国产亚洲精品一区二区www| 亚洲欧美一区二区三区黑人| av天堂中文字幕网| 麻豆国产av国片精品| 999久久久国产精品视频| 在线看三级毛片| 亚洲美女黄片视频| 欧洲精品卡2卡3卡4卡5卡区| 男女床上黄色一级片免费看| 中文字幕av在线有码专区| 国产精品女同一区二区软件 | 俺也久久电影网| 亚洲中文字幕一区二区三区有码在线看 | 老熟妇乱子伦视频在线观看| 男插女下体视频免费在线播放| av在线蜜桃| 午夜激情福利司机影院| www日本黄色视频网| 蜜桃久久精品国产亚洲av| 美女黄网站色视频| 亚洲成人久久性| 欧美乱妇无乱码| 我的老师免费观看完整版| 三级毛片av免费| 亚洲自拍偷在线| 欧洲精品卡2卡3卡4卡5卡区| 天堂动漫精品| 久久久久免费精品人妻一区二区| 亚洲aⅴ乱码一区二区在线播放| 一二三四在线观看免费中文在| 国产亚洲精品久久久com| 国产高清视频在线播放一区| 首页视频小说图片口味搜索| 日韩欧美一区二区三区在线观看| 免费在线观看日本一区| 我要搜黄色片| 日本成人三级电影网站| 国产美女午夜福利| www日本黄色视频网| 中文字幕最新亚洲高清| 久久精品综合一区二区三区| 一个人看视频在线观看www免费 | 两个人看的免费小视频| 久久精品综合一区二区三区| 免费在线观看视频国产中文字幕亚洲| 亚洲五月天丁香| av在线蜜桃| 99久久精品国产亚洲精品| 五月伊人婷婷丁香| 国产aⅴ精品一区二区三区波| 久久香蕉国产精品| 在线十欧美十亚洲十日本专区| 久久久久国内视频| 少妇的逼水好多| 亚洲九九香蕉| 国产精品98久久久久久宅男小说| 国产欧美日韩一区二区精品| 无遮挡黄片免费观看| 日本三级黄在线观看| 免费高清视频大片| 国产免费男女视频| 午夜免费激情av| 亚洲av成人不卡在线观看播放网| 国产精品久久电影中文字幕| 亚洲国产欧洲综合997久久,| 一本综合久久免费| 在线观看午夜福利视频| 亚洲av成人不卡在线观看播放网| 欧美中文综合在线视频| 国产黄a三级三级三级人| 又黄又粗又硬又大视频| 亚洲五月婷婷丁香| 免费看a级黄色片| 成人18禁在线播放| 成人特级av手机在线观看| 亚洲国产精品合色在线| 欧美xxxx黑人xx丫x性爽| 国产高潮美女av| 精品久久蜜臀av无| 天堂动漫精品| 免费在线观看视频国产中文字幕亚洲| 国产精华一区二区三区| 久久久久久人人人人人| 色av中文字幕| 变态另类丝袜制服| 日韩国内少妇激情av| 男插女下体视频免费在线播放| 国产精品久久久久久亚洲av鲁大| www.精华液| 国产黄a三级三级三级人| 免费看光身美女| 国产亚洲av嫩草精品影院| 女人高潮潮喷娇喘18禁视频| 亚洲欧洲精品一区二区精品久久久| 成人特级av手机在线观看| 白带黄色成豆腐渣| 国产精品美女特级片免费视频播放器 | 亚洲色图 男人天堂 中文字幕| 亚洲精品在线观看二区| 黄色丝袜av网址大全| 色综合站精品国产| 国产精品98久久久久久宅男小说| 国产麻豆成人av免费视频| 午夜a级毛片| 国产av麻豆久久久久久久| 99国产精品一区二区蜜桃av| 成人av一区二区三区在线看| 亚洲av电影不卡..在线观看| 国产黄a三级三级三级人| 国内精品一区二区在线观看| 一级作爱视频免费观看| 啦啦啦免费观看视频1| 日韩大尺度精品在线看网址| 51午夜福利影视在线观看| 日本黄大片高清| 老司机午夜福利在线观看视频| 深夜精品福利| 天天一区二区日本电影三级| 国产淫片久久久久久久久 | 天天一区二区日本电影三级| 亚洲色图 男人天堂 中文字幕| 亚洲国产精品sss在线观看| 九色国产91popny在线| 免费大片18禁| 亚洲一区二区三区色噜噜| 99久国产av精品| АⅤ资源中文在线天堂| 曰老女人黄片| 热99re8久久精品国产| 男人舔女人下体高潮全视频| 在线播放国产精品三级| 88av欧美| 久久这里只有精品19| 黄色丝袜av网址大全| 国产精品香港三级国产av潘金莲| 中文在线观看免费www的网站| 免费高清视频大片| 欧美乱码精品一区二区三区| 免费大片18禁| 老司机深夜福利视频在线观看| 男女视频在线观看网站免费| 欧美日韩国产亚洲二区| 91av网站免费观看| 老熟妇乱子伦视频在线观看| 亚洲五月婷婷丁香| 国产爱豆传媒在线观看| 成年免费大片在线观看| 久久精品aⅴ一区二区三区四区| 精品久久久久久久久久久久久| 亚洲精品色激情综合| 亚洲中文日韩欧美视频| 极品教师在线免费播放| 日本熟妇午夜| 99久久无色码亚洲精品果冻| 男人舔女人的私密视频| 亚洲天堂国产精品一区在线| 午夜亚洲福利在线播放| 我的老师免费观看完整版| 99热这里只有精品一区 | 日韩有码中文字幕| 精品国产美女av久久久久小说| 一本综合久久免费| 麻豆成人午夜福利视频| 天天躁狠狠躁夜夜躁狠狠躁| 国产精品电影一区二区三区| 麻豆av在线久日| 精品久久久久久,| 免费在线观看日本一区| 亚洲五月天丁香| 村上凉子中文字幕在线| 成人一区二区视频在线观看| 欧美xxxx黑人xx丫x性爽| 午夜福利18| 成人欧美大片| 99riav亚洲国产免费| 一级a爱片免费观看的视频| 日日摸夜夜添夜夜添小说| 亚洲精品久久国产高清桃花| 精品久久久久久,| 国产精品野战在线观看| 久久久久九九精品影院| 日韩 欧美 亚洲 中文字幕| 国产成人精品久久二区二区免费| 丝袜人妻中文字幕| 亚洲成人中文字幕在线播放| 亚洲中文日韩欧美视频| 禁无遮挡网站| 精品国产三级普通话版| 国产午夜精品论理片| 国产淫片久久久久久久久 | 全区人妻精品视频| 亚洲第一欧美日韩一区二区三区| 琪琪午夜伦伦电影理论片6080| 一区二区三区激情视频| 亚洲avbb在线观看|