• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Modified MMS:Minimization Approach for Model Subset Selection

    2023-12-12 15:50:30RajathiandRukmani
    Computers Materials&Continua 2023年10期

    C.Rajathi and P.Rukmani

    School of Computer Science and Engineering,Vellore Institute of Technology,Chennai,600127,India

    ABSTRACT Considering the recent developments in the digital environment,ensuring a higher level of security for networking systems is imperative.Many security approaches are being constantly developed to protect against evolving threats.An ensemble model for the intrusion classification system yielded promising results based on the knowledge of many prior studies.This research work aimed to create a more diverse and effective ensemble model.To this end,selected six classification models,Logistic Regression(LR),Naive Bayes(NB),K-Nearest Neighbor(KNN),Decision Tree (DT),Support Vector Machine (SVM),and Random Forest (RF) from existing study to run as independent models.Once the individual models were trained,a Correlation-Based Diversity Matrix (CDM)was created by determining their closeness.The models for the ensemble were chosen by the proposed Modified Minimization Approach for Model Subset Selection(Modified-MMS)from Lower triangular-CDM(L-CDM)as input.The proposed algorithm performance was assessed using the Network Security Laboratory—Knowledge Discovery in Databases(NSL-KDD)dataset,and several performance metrics,including accuracy,precision,recall,and F1-score.By selecting a diverse set of models,the proposed system enhances the performance of an ensemble by reducing overfitting and increasing prediction accuracy.The proposed work achieved an impressive accuracy of 99.26%,using only two classification models in an ensemble,which surpasses the performance of a larger ensemble that employs six classification models.

    KEYWORDS Ensemble learning;intrusion detection;minimization;model diversity

    1 Introduction

    Cybersecurity attacks have become increasingly complex and challenging to detect and prevent because of the rapid advancement of networks and related technologies[1].Organizations rely on various security measures,including firewalls,Intrusion Detection Systems(IDS),Intrusion Prevention Systems(IPS),anti-virus,anti-malware,two-factor authentication,encryption,vulnerability scanning,employee training,and awareness creation.However,IDS is often considered the best solution for intrusion detection owing to its ability of real-time monitoring,both signature and anomaly-based detection;it minimizes performance impact and helps organizations maintain their regulations.

    An IDS is an essential component of network security designed to detect and respond to unauthorized access and malicious activities.An IDS analyzes the network traffic or system log to identify patterns or behaviors that are consistent with known attacks or suspicious activities.Host-based IDS and Network-based IDS are the primary categories of IDS,where signature-based and anomalybased are the detection methods of IDS[2].Signature-based IDS depends on predefined patterns or signatures of known attacks and compares network traffic against these known signatures to detect potential threats [3].Anomaly-based IDS uses statistical or Machine Learning (ML) approaches to establish the baseline behavior of normal and deviations of the threats[2].

    Designing an effective IDS can be challenging owing to the diversity and complexity of modern threats and attacks.Employing a single classifier is not sufficient for constructing a powerful IDS[4].The ensemble approach has emerged as a promising approach to address these challenges and improve the performance of IDS [5].Ensemble learning for an IDS involves combining multiple models or classifiers to improve the overall accuracy and reduce false positives and false negatives [4].Three techniques are used to increase the performance of an IDS:bagging,boosting,and stacking.Bagging involves the use of multiple models on different subsets of data and combining their output,in boosting involves training a model sequentially and giving more weight to misclassified instances over time.Stacking involves combining the output of various models using meta-classifiers[6].

    Ensemble methods can be grouped into two major categories: homogenous and heterogeneous[4,7].Homogenous ensembles have the same types of models and architectures with different subsets of training data.Predictions of every subset are combined to produce the outcome.Bagging[3–6,8–12],Boosting[3–6,8,10–19]are the type of homogeneous ensembles,which uses different methods to determine final prediction.Specifically,majority voting[3–6,9–11,17,18],weighted voting[4,6,7,9,10],average voting are prevalent methods used in existing studies.Several models and architectures have been integrated to produce a final product in heterogeneous ensembles.These models employed different subsets of data,algorithms,and parameters.In a heterogeneous ensemble,stacking [3–6,8,14,18,20]is an example in which the output of various models is used as the input for a metamodel to integrate the results.Another type of model is the cascade,which uses the output of one model as an input to another.

    Various ensemble techniques have been implemented for intrusion detection in existing studies,although ensemble learning has limitations in certain aspects of detection.The main aim of the ensemble is to incorporate multiple models to improve performance in terms of accuracy,precision,recall,etc.Along with performance improvement,computational cost,maintenance,overfitting,transparency,and applicability are difficulties to be treated in ensemble learning.To address these issues,model selection for the ensemble is an important process because it determines the quality and diversity of individual models.The two main objectives of the model selection focused on the diversity and performance of the models.The diversity of the model is important for reducing overfitting and improving accuracy,robustness,and flexibility.Model diversity can be achieved by the measure’s correlation and entropy;this process is called ensemble pruning.Based on the correlation coefficients of the models,a subset of the models is generated for the ensemble.

    Ensemble models gained widespread adoption due to substantial improvements in accuracy and robustness.However,they encounter challenges arising from model complexity,resource requirements,maintenance,and training models.Existing studies have extensively covered different ensemble models,with a particular emphasis on feature selection.However,the selection of an appropriate model for an ensemble still raises uncertainties and concerns.The primary motivation behind the research is to address the crucial aspect of model selection in the context of ensemble methods,driven by the existing knowledge gap.This work aims to bridge the gap and provide insights into effective model selection techniques.

    The NSL-KDD dataset is commonly used in the field of network security and intrusion detection to evaluate model performance.It is essential to use datasets to assess the ability of models to generalize,compare their performances,and identify areas of improvement.The proposed model is evaluated using the NSL-KDD dataset and considered accuracy,precision,recall,and F1-score as evaluation metrics.In addition,ROC curves are used to summarize the performance of the model.

    The major contributions of this study are as follows:

    ? To develop an intrusion detection system with the classifiers LR,NBC,KNN,DT,SVM,and RF individually

    ? Generate a Correlation-Based Diversity Matrix(CDM)between individual classification algorithms to ensure the diversity of each model.

    ? To develop a Modified-Minimization approach for Model Subset Selection(Modified-MMS)to select the minimal model for the ensemble,that should achieve objective diversity and performance.

    ? To develop an aggregated prediction of model subsets using majority voting and compare the performance with all classifier ensemble combinations.

    2 Related Work

    The rapid growth of technology has increased the threat to network users[21].Providing privacy and mitigating security threats are important factors for multipurpose network applications [19].A vulnerability in a network causes significant damage to an organization [2],and a large amount of network traffic and a growing number of attacks are recognized as issues to deal with[22].Traditional detection methods lack the modality of data,and interdependence among features results in the model being unable to detect network attacks in real-time [14].The process of extracting important and relevant features from large amounts of network traffic data is a critical step in developing highly effective intrusion detection systems [23].The high False Positive Rate (FPR) and Detection Rate (DR) are the most common problems to be treated in IDS [10],which can be effectively improved by the ensemble methods compared to conventional single classification algorithms [12].In the ensemble process,moderately accurate components of classifiers are combined to obtain highly accurate classifiers [12].Several techniques have been proposed to produce efficient and diversified base classifiers for ensemble construction were discussed below.

    The effectiveness of the model is affected by redundant and irrelevant features.To eliminate pointless features from the dataset,the authors [21] suggested using an ensemble method with Recursive Feature Elimination (RFE),in which DT,SVM,and RF were used as base classifiers.To identify the irrelevant features of the attack groups Denial-of-Service (DoS),Probe,Remote to User (R2L),and User to Root(U2R),each classifier uses RFE to create an ensemble for prediction,to classify all types of attacks.The author[2]suggested a hybrid classification based on Ranker’s algorithm utilizing the Weka tool to address the weaknesses of network vulnerabilities.Ranker’s attribute assessment approach was used to select the pertinent attributes.The Pertinent attribute in the proposed strategy reduces the time complexity and resource utilization of the model.The classification algorithm selected for the proposed hybrid classification is Instance-Based Learner(IBk),Random Tree(RT),Reduced Error Pruning (REP) Tree,J48 graft,and RF.To construct a collection of ensemble solutions for successful intrusion detection,the author [11] presented a hybrid strategy that combines the Multi-Objective Genetic Algorithm(MOGA)and Neural Network(NN)to resolve the issue of bias toward attack classes,either majority or minority.The proposed work is implemented in two stages.In step 1,a set of Pareto optimum solutions is constructed,and each model response is combined using the majority vote.In step 2,the proposed strategy is compared against established bagging and boosting methods with performance metrics.To determine the best technique for intrusion detection,the authors[22]proposed a hybrid model using ML and Deep Learning(DL)techniques.The suggested hybrid strategy combines the pre-processing method with missing value management,data balancing with Synthetic Minority Oversampling Method (SMOTE),and Extreme Gradient Boosting(XGBoost)for the best feature selection to feed into algorithms to construct the model.ML models RF,DT,KNN,and DL models Multi-Layer Perceptron(MLP),Convolution Neural Network(CNN),and Artificial Neural Network(ANN)are used to develop a hybrid proposed approach and the suggested approach reduces overfitting,Type 1,and Type 2 error.

    To better serve applications,the authors[19]suggested an Optimized Ensemble-Intrusion Detection system(OE-IDS)using AutoML.The techniques utilized for data balance in the pre-processing stage include SMOTE,SMOTE-TOMEK(pair of instances of opposite classes in proximity),Adaptive Synthetic Sampling Approach(ADASYN),and Random Oversampling Examples(ROSE).Kendall’s test,Pearson’s correlation,and Spearman’s correlation were used for feature analysis.Soft voting was used to obtain the ensemble prediction results,and the performance was assessed using classification evaluation metrics.

    The most frequent issues in IDS that require treatment are high FPR and Detection Rate(DR).To address this issue,the authors[10]presented a two-stage model solution in which stage 1 offers a Pareto optimal solution for base models,and stage 2 makes the final prediction via a majority vote.Pareto-optimal solutions are produced using the Archived-based Micro Genetic Algorithm(AMGA)algorithm.The MLP classifier is the foundational model and the weights are initialized randomly.The best-performing models were chosen for an ensemble using overproduce and to choose the technique.The authors [4] suggested a dual ensemble,which combines two existing ensemble approaches,to decrease the False Alarm Rate (FAR) and boost DR.The fundamental learning algorithms are Gradient Boosting Decision Tree (GBDT) and bagging.The Classification and Regression Tree(CART) was regarded as a weak learner by the basic GBDT.The performance of the base learners was improved using a variety of dual ensembles,including the improved GBDT,Gradient Boosting Machine (GBM),LightGBM,Category Boosting (CatBoost),and XGBoost.The Friedman rank test was applied to rank ensemble performance.The disadvantages of traditional intrusion detection systems include their poor accuracy and detection rate,the authors[12]proposed Ensemble Learning algorithm-based Anomaly Detection in Communication Networks(EL-ADCNS)is a unique anomaly detection technique that is suggested as a solution to the problem.Correlation-based Feature Selection(CFS)and RF are used to lower the dimensionality.RF,SVM,AdaBoost,and Bagging classifiers are used in hybrid Ensemble Learning(EL)training,and the result is average-voted before being input to the Hybrid Ada Boosting Bagging Algorithm(HABBA)classifier[12].

    To address the key difficulty for IDS and IPS in malware prediction in new locations with high accuracy and detection rate,the authors[15]proposed the ensemble’s implementation using accuracy and Kappa value in the Weka tool.The foundation classifiers for ensembles are NB,LR,and DT,whereas the aggregation techniques employed are Bagging and Boosting.To defend the computing infrastructure against the metamorphic and polymorphic behavior of malware,a powerful IDS is required.The XGBoost,Bagging Classifier,Extra Tree,and RF classification algorithms are the learners for the proposed[3]Stacked Ensemble-based IDS(SE-IDS).The output of the base learner is provided to the meta-learner as input for the final classification;MLP is used here.Various sets of features are sent to each base learner,and each model’s final value is determined by majority voting before being fed to the meta-learner.A varied set of features was chosen using Sequential Forward feature Selection(SFS).Ten separate runs were conducted to test the performance of the model.

    For the development of new attacks,a single classification system and an outdated dataset are insufficient.For dimensionality reduction,the authors[5]presented a heuristic technique based on the CFS and Bat Algorithm(BA).The basis classifiers for the ensemble consist of C4.5,RF,and Forest by Penalizing Attribute(Forest PA).Soft voting,which uses the average probability,was used to make the final prediction.To address new threats to network-connected systems,the authors[18]presented an ensemble of discriminant classifiers.A discriminant classifier is a method for transforming a weak learner into a strong learner.The model selection for the ensemble was performed using a random subspace approach.Empirical analysis of the 5-cross-fold subgroup that was chosen at random.Here,four were utilized for training,one for testing,and the final prediction was made by majority vote.

    In the ensemble process,moderately accurate components of classifiers are combined to obtain highly accurate classifiers.The author [7] proposed a bagging and arcing (adaptively resample and combine)ensemble with homogenous and heterogeneous models respectively,and the performance of the models was evaluated using accuracy measures.The Radial Basis Function(RBF)and SVM are the classifiers used as a base classifier,and a 10-fold cross-validation technique is applied to calculate accuracy.Bagging was performed with RBF and SVM and compared with the base classifiers in terms of accuracy.For heterogeneous ensembles using arcing classifier creates a hybrid RBF-SVM and the final output is decided by weighted voting.

    Multiple M-AdaBoost with Particle Swarm Optimization (PSO) and Modified AdaBoost with an area under the curve (M-AdaBoost-A) were proposed by the authors [13] to create an effective intrusion detection system to deal with large-scale,unbalanced,and multiclass data.The base learner,sub-learner,and expert learner in the M-AdaBoost-A-SMV model are integrated using simple majority voting.

    Conventional detection techniques fall short of the real-time network dataset requirements.The authors[14]introduced Multi-dimensional Feature Fusion and Stacking Ensemble Mechanism(MFFSEM),which effectively detects anomalous behavior from real-time network data.Map Reducebased Ensemble for IDS(MR-EIDS)was proposed[20]to detect intruders and attackers from realtime datasets.The security model is a key component of the fog-computing network,which facilitates the quality of service.The authors [24] suggested a hybrid optimization-driven ensemble classifier using the Ride Sea Lion Optimization (RSLO) algorithm,which combines the Rider Optimization Algorithm (ROA) and Sea Lion Optimization Algorithm (SLnOA).For feature selection,a filter approach based on the Kolmogorov-Smirnov correlation was applied.The categorization models utilized in the ensembles are the RideNN,Deep Neuro-Fuzzy Network (DNFN),and Shepard Convolution Neural Network(ShCNN).Comprehensive features were chosen from the fundamental characteristics to be combined with other features and input into the classifiers.By adding the probabilities of each classifier,a meta-classifier is created,which generates the final prediction from the classifiers.employing DT and RF to stack the classifiers for the ensembles.The combination of classes was evaluated along with the precision,recall,accuracy,and F1-score.Big data plays a major role in organizations that extract useful information from data,for which the authors[25]suggested an intelligent intrusion detection system employing binary Grey Wolf Optimization(GWO)to safeguard the power grid from intrusion and ensure its dependable and efficient operation.GWO is used for feature selection in Binary Grey Wolf Optimization Ensemble Classification(BGWO-EC).

    An excellent Internet-of-Things(IoT)infrastructure is required for the efficient operation of smart cities.The authors[8]suggested an ensemble method for intrusion detection to improve the security of IoT-based applications.The author employed classification models RF,KNN,ANN,and SVM as stacking ensemble techniques used for model aggregation.Providing security for an IoT-based global network.Local Search-Pigeon Inspired Optimization(LS-PIO)has been proposed[9]for the selection of features from datasets.The author used PIO,Tabu Search,and Hill Climbing as local search techniques,and weighted voting was used to determine the outcome.The ensemble combinations used were One-Class Support Vector Machine (OC-SVM),Isolation Forest (IF),and Local Outlier Factor(LOF)for effective classification.In addition,the author assessed the complexity of the phases of feature selection and intrusion detection.IoT combines sensors and devices to automate routine business processes in industries.By adopting a voting-based ensemble approach,the authors [16]proposed a framework for the Industrial IoT (IIoT) to identify cyberattacks.Histogram gradient Boosting(HGB),CatBoost,RF,and the final prediction made using the hard voting method made up the ensemble.The author suggests a two-step approach that entails traffic analysis and abnormal event detection to address the security issues with IoT systems.The ensemble containing Extra tree,RF,and Deep Neural Network(DNN)was utilized for the traffic analysis step and the event detection step,respectively.Accuracy,precision,recall,F1-score,and Balanced Accuracy(BAcc)are the performance metrics utilized for evaluation.BAcc measures the detection performance in an unbalanced dataset by averaging the recall for each class[26].SVM integration with Chaos Game Optimization(CGO)was suggested by the authors[17]to handle heterogeneous data.The final prediction made by voting is included in the k times SVM(SVM1,SVM2,...,SVMk)used as an ensemble for training.

    In the existing studies on intrusion detection,numerous classification models have been explored.But,many of these models have focused widely on features selection and performance improvement of the ensemble model.From the existing studies,it is observed that there is a lack of transparency in the process of model selection for an ensemble.This lack of clarity hinders the ability to justify the choices for constructing an ensemble.The table presented below showcases a compilation of diverse ensembles derived from existing studies,as depicted in Table 1.

    Table 1:Combination of ensembles used in existing work

    The selection of a dataset is an important component in building an efficient machine-learning model.According to the State of Data Science 2022 report [27],preparation and understanding of the data is an important and time-consuming task of ML model building.To serve a real-world application,model building should be evaluated using quality data.Depending on the need and problem statement the authors used various datasets for the intrusion detection system.The datasets listed in Table 2 have been widely used and proven to be valuable resources in the existing literature on intrusion detection.

    Table 2:Various datasets used for intrusion detection in existing work

    3 Proposed Work

    The proposed work objective is to address the model selection for an ensemble by selecting the most suitable models based on the application requirements.Through the implementation of the proposed Modified MMS approach,the aim is to offer a justification for model selection within an ensemble.Furthermore,research findings indicate that the identified models can be combined in an ensemble that can reduce the space complexity of the problem.In the following section,the steps involved in the proposed approach are described.

    3.1 Model Selection for Ensemble

    Ensemble learning combines the predictions of various models to produce a powerful and reliable model that can be generalized more effectively to new data.Because it specifies which model will be included in the ensemble and how it will be mixed,model selection is a crucial stage in ensemble learning.The caliber of the individual models and the ensemble diversity both affect the performance of the ensemble.Selecting models that are complementary to one another and that have various strengths and weaknesses is crucial.Fig.1 depicts the overall architecture of the proposed system which comprises various modules,including preprocessing,individual model training,creating a correlation matrix for the individual models,selection of the models for the ensemble,and aggregation.

    Figure 1:Proposed method architecture.A model which comprises dataset,pre-processing,and the proposed model selection Modified-MMS modules.The models selected from the proposed system are aggregated to produce the final class prediction

    3.1.1 Correlation-Based Diversity Matrix(CDM)

    Model selection for an ensemble relies on the objective of improving the performance and diversity of the model.The importance of diversity in ensemble learning lies in the fact that it helps reduce overfitting and improves model robustness.CDM is used to measure the diversity between individual models in an ensemble.It evaluates the similarity between different models in terms of their predictions using the same input data.

    To create CDM constructed as follows:

    Step 1–For each model,generate a set of predictions

    Step 2–Calculate the correlation coefficients of all the individual models

    Step 3–Create a Correlation-based Diversity Matrix.

    The linear relationship between the variables is defined by the statistical measure correlation coefficient,which also describes the degree of the relationship between two variables.Here the variables represent the predictions generated by each model.The correlation coefficient ranges from -1 to 1,where a value of-1 indicates a negative correlation,0 indicates no correlation and 1 indicates a positive correlation.

    The CDM can be visualized in a square matrix,where each row and column represent a model in an ensemble.The diagonal of the matrix contains the correlation coefficient between the model and itself,which is always equal to one.The off-diagonal elements contain the correlation coefficients between the pairs of models.From the pairs of models select the models,with low correlation coefficients to maximize the performance and diversity of the ensemble.Highly correlated models mean that they are making a similar error and may not provide much new information to an ensemble.

    If the ensemble has annclassifier,then the matrix will ben×n.

    Here,clf ij,indicates the classifier row,and column indexes represented byiandj,respectively.

    3.1.2 Exploiting Symmetry of Matrix

    A symmetric matrix is a matrix in which the transpose of the given matrix is equal to the original matrix.In a symmetric matrix,the entries above the diagonal are equal to those below the diagonal.This means that matrix is completely determined by its diagonal entries and either the upper or lower triangular parts.By taking advantage of symmetry,computational complexity,and memory usage can be reduced.

    Let us consider the CDM from Eq.(1),if the transpose of CDM is equal to the original CDM,then this implies that the matrix is symmetrical about its diagonal.This can be expressed mathematically as follows:

    In terms of the matrix elements,the symmetry property can be expressed as:

    where a_ij is the element of the ith row and jth column of the matrix.

    If the given matrix satisfies the property of symmetry,either the lower triangular part or the upper triangular part can be used further,which provides the advantage of computational efficiency and reduces storage requirements.

    3.1.3 Modified Minimization Approach for Model Subset Selection(Modified-MMS)

    Minimization is the process of finding the smallest possible value of a quantity,typically a function.From the CDM,the lower triangular part(L-CDM)is considered for subset selection.From the L-CDM the model subsets must be selected for input to the ensemble.Subset for the ensembles selected from the following steps.

    Setting up the threshold significantly impacts the performance of the classification model.To improve the diversity of the model,a combination of low-correlated models works efficiently.Combining the models with a high correlation will not be able to predict new errors and improve the accuracy.

    To select subsets for an ensemble,this research opted to use the concept of Deterministic Finite Automata(DFA)minimization.DFA minimization is the process of reducing the number of states in DFA while preserving its language recognition capability.The Myhill-Nerode Theorem or Table filling method[28]is used in the minimization of DFA,in which the minimal equivalent version of any DFA includes a minimum number of states possible.

    Steps for minimization in DFA[29]:

    1.Create a pair of states involved in a given DFA.

    2.Mark all pairs(Qa,Qb)such that Qais the final state and Qbis a non-final state.

    3.If there is any unmarked pair(Qa,Qb)such thatδ(Qa,x)andδ(Qb,x)are marked,then mark(Qa,Qb).where x is an input symbol.Repeat this step until no more marking can be made.

    4.Combine all the unmarked pairs and make them a single state in minimized DFA.

    The proposed Modified-MMS uses the concept of table filling method of DFA Minimization.In the Modified-MMS the base classifiers are considered nodes,and the Threshold T (Correlation Coefficient Value) should have a minimum correlation value.A node that has low performance in terms of evaluation metrics is considered a threshold node.Fig.2 shows the workflow of the proposed Modified MMS.

    Step 1–Let M1,M2,...,Mn be a set of models

    Create a pair with all states(models)from L-CDM.

    Step 2–Let C1,C2,...,Cm be a set of conditions

    Step 3–Combine the pairs Clfiand Clfjas a subset to input ensemble,herei,jindicates the models.

    If value[(Clfi,Clfj)]<=T,then Clfiand Clfj←Sk(Subset k)

    Combine the pairs Clfi and Clfj as a subset to input ensemble

    If(Clfi€threshold node)OR(Clfj€threshold node),then Clfiand Clfj←Sk(Subset k)

    Step 4–Make the pairs to input.

    Let X be a set of n classifiers,and let r(i,j)be the correlation coefficient between i and j.Then the subset S of the models that have pairwise correlation coefficient less than or equal to thresholdtcan be defined as:

    where x_i is the i-th model in X,and the condition for inclusion in the subset is:

    The subset S includes all classifiers x_i in X such that the correlation coefficient between x_i and every other variable x_j in X is less than or equal to thresholdt.

    Figure 2:Workflow representation of Modified-MMS

    Algorithm:Modified-MMS

    3.2 Aggregation Method

    An aggregation method is a technical approach to integrating the predictions generated by multiple individual models.Each model is trained on a subset of the data and then makes its prediction.These individual predictions are aggregated using specific voting rules to produce the final predictions.The most used aggregation methods in the existing study include majority voting,weighted voting,and stacked voting.The voting aims to improve the accuracy and robustness of the models,especially when individual models are prone to overfitting or have a high bias.

    Subsets k created in the phase Modified-MMS trained with the dataset individually and the result was obtained using majority voting by combining the prediction.The implementation of majority voting serves to enhance both the accuracy and robustness of the model by effectively minimizing the influence of errors and biases associated with individual models.

    Let us consider N individual models in an ensemble,each of which produces a binary outcome of 0 or 1.Let x_i be the output of the ith model,where i=1,2,...,N.

    The majority voting combines the outputs of all N models to produce the final prediction(0 or 1).

    For example,consider 5 classifiers in ensemble,and the prediction of each model is

    Then,the final prediction is

    sum(x_i)=3,which is greater than 5/2,so the final prediction is x=1.

    4 Experimental Setup

    This section outlines the key components of the experiment,including the dataset description and validation procedure employed.

    4.1 Data Set Description

    The NSL-KDD dataset is an effective benchmark dataset that is widely used to compare the performance of models in IDS,which is a refined version of KDD CUP’99.Furthermore,the effectiveness of the dataset can be improved by default partitioning to KDD_Train+,KDD_Test+,and KDD_Train+_20 percent.The advantage of partitioning is that it runs the model conveniently and efficiently.A subset of NSL-KDD 20 percent data evaluates the performance of intrusion detection and machine learning model more efficiently and requires minimal resources than using the entire NSL-KDD dataset.

    The dataset used for an experiment contains 25192 instances,in which 13449 are normal and 11743 are abnormal connections.Traffic input for intrusion detection referring 41 features and the target column represent the label‘normal’or‘a(chǎn)bnormal’.The dataset is publicly available to download,along with the full NSL-KDD dataset and documentation.The initial pre-processing steps involve several tasks,including identifying and handling missing and duplicate values,performing encoding,scaling the features,and splitting the data into training and testing sets,all of which are necessary to prepare the data to load into the model.

    4.2 Evaluation Parameters

    The performance of the proposed models was evaluated using the following performance metrics:Accuracy,precision,recall,and F1-score[30].Defining evaluation metrics for a model is an integral part of estimating the accuracy of future data.To derive performance measures,an understanding of the confusion matrix is required.The confusion matrix is a representation of the prediction results.Each prediction can be any one of the outcomes based on how equivalent it is to the actual value.

    4.2.1 Accuracy

    It is a measure of the performance of the classification model,which indicates the proportion of correct predictions made by the model with overall prediction.

    Accuracy is a suitable measure for dealing with balanced datasets.The precision,recall and F1-score are more suitable when the data are imbalanced [31].Accuracy is the measure of correctly classified instances;similarly,misclassification rate is the measure of incorrectly classified instances.

    4.2.2 Precision

    The proportion of correctly predicted positive classes and total positive observation

    4.2.3 Recall

    The proportion between correctly predicted positive class and actual positive observations

    4.2.4 F1-Score

    The F1-score was the harmonic mean of the precision and recall system measurements.The overall correctness of the model was derived from the F1-score.

    5 Results and Analysis

    Following the experimental setup,this section presents the obtained results and a comprehensive analysis of the performance of the proposed Modified MMS.

    5.1 Correlation-Based Diversity Matrix(CDM)

    In conducted experiment,six base classifiers were utilized for training: LR,NBC,KNN,DT,SVM,and RF.Each classifier was assigned a label:clf1,clf2,clf3,clf4,clf5,andclf6,respectively.The correlation between the predictions of these individual classifiers was represented as a matrix denoted by CDM and presented in Eq.(1).

    Eq.(13)is derived from Eq.(1),which represents the correlation between the prediction of the six base classifiers used in the experiment.In the matrix representation,each row and column in CDM corresponds to a specific base classifier,while the value in each cell of the matrix indicates the degree of correlation between the predictions of the pair of classifiers.By analyzing CDM insights model relationships and predictive performance can be derived.

    The CDM can be effectively visualized as a grid,which is shown in Fig.3.Each row and column in the grid correspond to an individual base classifier used in experiments.To make the matrix more interpretable the values are represented as a heat map,where darker colors indicate strong correlation and lighter colors indicate weak correlation.

    Figure 3:Heat map of base classifier correlation

    From Eq.(2),the property of matrix symmetry is utilized to significantly reduce the computational cost and the memory requirement for certain operations.The correlation between two classifiers is the same regardless of the order in which they are considered and can utilize the advantage of symmetry property.The transposition of the CDM is shown in below Eq.(14):

    From Eq.(14),either the lower or upper half part of the matrix represents the correlation values,which eases further implementation.For a conducted experiment on model subset selection,the lower triangular part of the matrix is chosen,denoted as L-CDM.Considering the correlation values for the pairs of variables where the row index is greater than the column index.The L-CDM is represented in Eq.(15)below:

    Model subsets from L-CDM(2,1),(3,1),(3,2),(4,1),(4,2),(4,3),(5,1),(5,2),(5,3),(5,4),(6,1),(6,2),(6,3),(6,4),(6,5)were chosen for the next phase of minimization.

    5.2 Modified Minimization Approach for Model Subset Selection(Modified-MMS)

    DFA minimization aims to minimize the states in an automaton,in which the minimized DFA is also accepted by the language as the original DFA.Acquiring the concept of minimization to select the minimal model which should produce the same performance of multiple model combinations.Selecting the minimal models for an ensemble reduces the complexity,enhanced diversity,and improved efficiency.

    Here the classifiersclf1,clf2,clf3,clf4,clf5,and clf6are considered as a state of automaton,the correlation coefficient between the models is considered as a transition value,and the low-performing model/classifier consider as an end node in an automaton.The subset derived from L-CDM is given to the proposed Modified-MMS method to minimize the ensemble models.The steps involved in the model selection are itemized below.

    In the first step,create pairs of all the models derived from L-CDM.These pairs include (2,1),(3,1),(3,2),(4,1),(4,2),(4,3),(5,1),(5,2),(5,3),(5,4),(6,1),(6,2),(6,3),(6,4),(6,5).Then group these models according to their components.The LR model(denoted byclf1)has combinations(2,1),(3,1),(4,1),(5,1),and(6,1)as depicted in Fig.4a.This work selects the subset of combinations that satisfy the Modified-MMS condition for the ensemble input.Similarly,repeat this process for all the base classifiers,which include NBC,KNN,DT,SVM,and RF (denoted asclf2,clf3,clf4,clf5,andclf6,respectively).The classifiers combination NBC,KNN,DT,SVM,and RF are depicted in Figs.4b–4f.

    Figure 4:Creating subsets concerning State i (i=clf1,clf2,clf3,clf4,clf5,clf6),clfi represents classification models

    Based on Fig.4.the selected model combinations that satisfy the condition for the ensemble input are as follows:

    ? From Fig.4a.the subset selected is(2,1),which indicates that the recommended combination for an ensemble is NBC and LR.

    ? From Fig.4b.the selected subsets are(3,2),(4,2),(5,2),and(6,2)which include the combinations (KNN,NBC),(DT,NBC),(SVM,NBC),(RF,NBC) respectively.Other combinations failed to satisfy the rules of Modified-MMS.

    The above-mentioned model combinations,namely S1(NBC and LR),S2(KNN and NBC),S3(DT and NBC),S4 (SVM and NBC),and S5 (RF and NBC) are the recommended ensemble from Modified-MMS.

    Table 3 displays the performance of the base classification models,while Table 4 shows the combination of a subset of these models selected using the proposed Modified-MMS technique and their corresponding performance.The hyperparameter for the selected classifiers was calculated using grid search cv=5,and the evaluated performance was recorded in Table 5.The results from Tables 4 and 6 indicate the noticeable improvement in accuracy,precision,recall,and F1-score of the ensembles.Additionally,the performance of the model subsets compared with the ensemble made with all base classification models.The proposed research work achieved higher performance with only two base classifiers than the six-based classifier ensemble.

    Table 3:Performance of base classification models

    Table 4:Performance of selected models from proposed Modified-MMS for ensemble

    Table 5:Selected hyperparameter using grid search CV

    Table 6:Performance of selected models from proposed Modified-MMS for ensemble with hyperparameter tuning

    Hyperparameter tuning is the process of selecting an optimal set of hyperparameters for the model.Selecting the hyperparameter for the model is a crucial part to achieve high performance.In conducted experiment Grid Search CV (Cross-Validation) to find the best hyperparameter for the model is used.The selected hyperparameter for the experiment data is listed below in Table 5.

    The term ROC stands for Receiver Operating Characteristic,and it is commonly used to evaluate the effectiveness of a binary classification system.The ROC curve is created by plotting True Positive Rate(TPR)against FPR for various classification thresholds.By analyzing the ROC curve,a performance summary of the classifier is obtained.From the results of the conducted experiment,Fig.5 illustrates the performance of each classifier,while Fig.6 demonstrates the performance of the base classifier and the ensemble of all base classifiers.From Fig.7,the performance of each subset selected from the proposed Modified MMS is identified.Fig.8 shows the comparison of different performance metrics of selected subsets.

    Figure 5:ROC between all base classifiers

    Figure 6:ROC between base classifier vs.ensemble of classifier

    Figure 7:Comparison of ROC curves between selected subsets from proposed Modified-MMS

    Figure 8:Comparative analysis of individual models and their ensemble subsets

    6 Conclusion and Future Work

    In this study,the random forest algorithm exhibited strong performance as a standalone classification model,achieving an accuracy of 99.02%.However,to further enhance the performance and diversity of the model,a Modified-MMS technique was proposed to select a subset of the ensemble models.The proposed system selected only two models for the subset,which resulted in substantial improvements in performance and diversity as evaluated by metrics.Interestingly,the subset ensembles outperformed the ensembles of all six base classifiers.The subset containing only Random Forest and Naive Bayes classifiers (Subset 5) from Table 5 achieved an accuracy of 99.26%,surpassing the ensemble of all six base classifiers which achieved an accuracy of only 98.31%refer to Table 4.Subset 4 consisting of a Support Vector Machine and Naive Bayes classifiers achieved a precision score of 99.47%,Subset 2 with K-Nearest Neighbors and Naive Bayes classifiers achieved a recall score of 99.43%,and Subset 5 with Random Forest and Naive Bayes classifiers provided an F1-score of 99.30%.All these scores were higher than the corresponding metrics obtained from the ensemble of all six base classifiers.This indicates that the combination of just two base classifiers can achieve both high performance and diversity,as compared to the ensemble of all six classifiers.

    The task of selecting an appropriate model for an ensemble is critical as it can greatly influence the ensemble’s performance.Incorporating models that have varying characteristics can lead to a substantial improvement in performance,as opposed to combining similar models.To address this objective,a novel approach called Modified-MMS for selecting a subset for an ensemble is proposed.Apart from advantages,the proposed minimization approach suffers from the non-uniqueness problem,where a unique solution for the model selection cannot be fixed.The process of combining models for the ensemble relies on factors such as the choice of base classifiers,the selection of performance metrics,the diversity of models,and the specific problem at hand.

    Acknowledgement:We would like to thank the support of School of Computer Science and Engineering(SCOPE),VIT,Chennai.

    Funding Statement:The APC was funded by the Vellore Institute of Technology(VIT).

    Author Contributions:Study conception and design:Rukmani P,Rajathi C;data collection:Rajathi C;analysis and interpretation of results:Rukmani P,Rajathi C;draft manuscript preparation:Rajathi C.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data presented in this study are openly available online:https://www.kaggle.com/datasets/hassan06/nslkdd?select=KDDTrain%2B_20Percent.txt.Accessed on 13 January 2023.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    中文字幕色久视频| 国产精品久久久久久精品电影小说| 国产视频一区二区在线看| 十八禁高潮呻吟视频| 成年人免费黄色播放视频| 日韩人妻精品一区2区三区| 国产成人91sexporn| av在线播放精品| 99热全是精品| 成人午夜精彩视频在线观看| 国产成人精品久久久久久| 久久99精品国语久久久| 99国产精品一区二区蜜桃av | e午夜精品久久久久久久| 女人高潮潮喷娇喘18禁视频| 国产精品99久久99久久久不卡| 国产老妇伦熟女老妇高清| 精品少妇内射三级| 9色porny在线观看| 久久ye,这里只有精品| 可以免费在线观看a视频的电影网站| 99热国产这里只有精品6| 超色免费av| 亚洲国产精品一区二区三区在线| 色94色欧美一区二区| 夫妻性生交免费视频一级片| 国产成人精品久久二区二区91| 操美女的视频在线观看| 精品高清国产在线一区| 精品人妻一区二区三区麻豆| 国产高清不卡午夜福利| 一级毛片 在线播放| 国产成人a∨麻豆精品| 亚洲精品国产区一区二| 人妻一区二区av| 免费看十八禁软件| 中文字幕最新亚洲高清| 热re99久久精品国产66热6| h视频一区二区三区| 丝袜喷水一区| 午夜福利一区二区在线看| av视频免费观看在线观看| 日韩一卡2卡3卡4卡2021年| 又大又黄又爽视频免费| 99热网站在线观看| 国产片内射在线| 少妇猛男粗大的猛烈进出视频| 老司机亚洲免费影院| 亚洲欧美一区二区三区国产| 久久国产精品影院| 久热爱精品视频在线9| 亚洲午夜精品一区,二区,三区| 在线av久久热| 男人操女人黄网站| 麻豆av在线久日| 亚洲av日韩在线播放| 巨乳人妻的诱惑在线观看| 好男人电影高清在线观看| 建设人人有责人人尽责人人享有的| 婷婷色综合www| 激情五月婷婷亚洲| 婷婷色综合大香蕉| 最新在线观看一区二区三区 | 极品人妻少妇av视频| 亚洲欧洲国产日韩| 国产老妇伦熟女老妇高清| 性色av乱码一区二区三区2| 国产三级黄色录像| 免费黄频网站在线观看国产| 午夜激情av网站| 亚洲熟女毛片儿| 中文字幕高清在线视频| 亚洲一区中文字幕在线| av不卡在线播放| 中文字幕人妻丝袜一区二区| 韩国精品一区二区三区| 欧美日韩亚洲高清精品| 天堂俺去俺来也www色官网| 激情视频va一区二区三区| 日韩中文字幕欧美一区二区 | 欧美老熟妇乱子伦牲交| 男女无遮挡免费网站观看| 人人妻,人人澡人人爽秒播 | 午夜福利在线免费观看网站| 色94色欧美一区二区| 久久人妻熟女aⅴ| 亚洲五月婷婷丁香| 一级毛片电影观看| 久久国产精品影院| 欧美日本中文国产一区发布| 少妇的丰满在线观看| 黄色a级毛片大全视频| 欧美国产精品va在线观看不卡| 老司机亚洲免费影院| 热99国产精品久久久久久7| 成年人午夜在线观看视频| 国产片特级美女逼逼视频| 中文字幕av电影在线播放| 制服人妻中文乱码| 久久国产精品人妻蜜桃| 久久久久久久久免费视频了| 国产免费一区二区三区四区乱码| 精品国产国语对白av| 女性被躁到高潮视频| 国产黄色视频一区二区在线观看| 国产一级毛片在线| 自线自在国产av| 国产又色又爽无遮挡免| 免费高清在线观看日韩| 一级片'在线观看视频| 亚洲欧美中文字幕日韩二区| 蜜桃在线观看..| 久久性视频一级片| 日日摸夜夜添夜夜爱| 亚洲av成人不卡在线观看播放网 | 精品国产一区二区久久| 在现免费观看毛片| 国产成人一区二区在线| 亚洲欧洲国产日韩| 国产精品一区二区精品视频观看| 精品人妻熟女毛片av久久网站| 亚洲天堂av无毛| 少妇猛男粗大的猛烈进出视频| 国产日韩欧美亚洲二区| 久久天躁狠狠躁夜夜2o2o | 国产高清不卡午夜福利| xxx大片免费视频| 欧美日韩av久久| 国产精品国产av在线观看| 亚洲五月婷婷丁香| 免费一级毛片在线播放高清视频 | 精品国产超薄肉色丝袜足j| 岛国毛片在线播放| 啦啦啦在线观看免费高清www| 国产精品99久久99久久久不卡| 人妻一区二区av| 黄色片一级片一级黄色片| 一本大道久久a久久精品| 日韩精品免费视频一区二区三区| 一区在线观看完整版| 久久国产精品男人的天堂亚洲| 尾随美女入室| 亚洲伊人久久精品综合| 久久久久网色| 国产三级黄色录像| 国产精品 欧美亚洲| 一级黄色大片毛片| 午夜福利免费观看在线| 欧美国产精品一级二级三级| 中文字幕亚洲精品专区| 高清黄色对白视频在线免费看| 夫妻午夜视频| 人人妻,人人澡人人爽秒播 | 欧美精品啪啪一区二区三区 | 99精国产麻豆久久婷婷| 国产视频首页在线观看| 精品亚洲成国产av| 青草久久国产| 亚洲自偷自拍图片 自拍| 亚洲国产毛片av蜜桃av| 成人黄色视频免费在线看| 高潮久久久久久久久久久不卡| 精品少妇久久久久久888优播| 丰满迷人的少妇在线观看| 久久99热这里只频精品6学生| 国产精品久久久久久精品电影小说| 亚洲精品国产av蜜桃| 国产免费又黄又爽又色| 十八禁网站网址无遮挡| 久久精品国产亚洲av高清一级| 久9热在线精品视频| 中文字幕色久视频| 亚洲精品av麻豆狂野| 亚洲欧美激情在线| 亚洲成人免费av在线播放| 国产av一区二区精品久久| 亚洲中文av在线| 国产一区二区三区av在线| 高清av免费在线| 免费日韩欧美在线观看| 男女无遮挡免费网站观看| 制服诱惑二区| 女性被躁到高潮视频| 91精品三级在线观看| 18禁裸乳无遮挡动漫免费视频| 一级a爱视频在线免费观看| 99re6热这里在线精品视频| 色婷婷av一区二区三区视频| 婷婷成人精品国产| 色精品久久人妻99蜜桃| 热re99久久国产66热| 亚洲国产毛片av蜜桃av| bbb黄色大片| 欧美精品啪啪一区二区三区 | 精品国产一区二区三区久久久樱花| 欧美另类一区| 国产亚洲精品第一综合不卡| 制服诱惑二区| 一级a爱视频在线免费观看| 五月开心婷婷网| 啦啦啦在线观看免费高清www| 蜜桃国产av成人99| 亚洲欧美一区二区三区黑人| 久久影院123| 婷婷成人精品国产| 一二三四社区在线视频社区8| 国产一区二区激情短视频 | 久久人人爽人人片av| 1024香蕉在线观看| 免费在线观看黄色视频的| 天天躁日日躁夜夜躁夜夜| 亚洲av在线观看美女高潮| 人体艺术视频欧美日本| 首页视频小说图片口味搜索 | av网站免费在线观看视频| av天堂久久9| 建设人人有责人人尽责人人享有的| 国产欧美亚洲国产| 性少妇av在线| 九色亚洲精品在线播放| 亚洲成人国产一区在线观看 | 亚洲成人免费电影在线观看 | 久久久精品国产亚洲av高清涩受| 精品一区在线观看国产| 黄色视频在线播放观看不卡| 精品人妻1区二区| 国产熟女欧美一区二区| 超色免费av| 成人18禁高潮啪啪吃奶动态图| 亚洲精品久久午夜乱码| 亚洲欧美日韩另类电影网站| 国产又爽黄色视频| 老司机影院毛片| 欧美日韩福利视频一区二区| 9色porny在线观看| 在线观看www视频免费| 两个人免费观看高清视频| 久久精品亚洲av国产电影网| 午夜免费鲁丝| 午夜激情av网站| 国产亚洲精品第一综合不卡| 成年美女黄网站色视频大全免费| 国产欧美日韩综合在线一区二区| 欧美大码av| 精品久久久精品久久久| 黑人巨大精品欧美一区二区蜜桃| 国产精品香港三级国产av潘金莲 | 咕卡用的链子| 99久久人妻综合| 99re6热这里在线精品视频| 婷婷色麻豆天堂久久| 成人三级做爰电影| 亚洲精品国产av成人精品| 午夜影院在线不卡| 高清黄色对白视频在线免费看| 男女国产视频网站| 国产精品一二三区在线看| 亚洲欧美一区二区三区黑人| 人妻人人澡人人爽人人| 亚洲精品美女久久久久99蜜臀 | 精品少妇黑人巨大在线播放| 欧美少妇被猛烈插入视频| 丝瓜视频免费看黄片| 久久性视频一级片| 国产亚洲av高清不卡| 一级毛片黄色毛片免费观看视频| 美女福利国产在线| 熟女av电影| 国产精品人妻久久久影院| 18禁观看日本| 最黄视频免费看| 亚洲欧美一区二区三区国产| 国产国语露脸激情在线看| 狂野欧美激情性bbbbbb| 国产av精品麻豆| 精品一区二区三区av网在线观看 | 亚洲国产欧美网| 亚洲成av片中文字幕在线观看| 欧美中文综合在线视频| av天堂在线播放| 大香蕉久久成人网| 亚洲精品一卡2卡三卡4卡5卡 | 嫁个100分男人电影在线观看 | 69精品国产乱码久久久| 国产无遮挡羞羞视频在线观看| 大话2 男鬼变身卡| 精品久久久精品久久久| 黑丝袜美女国产一区| av国产精品久久久久影院| 两个人免费观看高清视频| 久久久久视频综合| 蜜桃国产av成人99| 如日韩欧美国产精品一区二区三区| www日本在线高清视频| 九草在线视频观看| 日日夜夜操网爽| 免费看不卡的av| 91麻豆av在线| 国产又爽黄色视频| 91麻豆av在线| 日本欧美视频一区| 少妇的丰满在线观看| 久久精品国产a三级三级三级| 国产色视频综合| 97精品久久久久久久久久精品| 你懂的网址亚洲精品在线观看| 国产成人精品久久二区二区免费| 黄色一级大片看看| 脱女人内裤的视频| 1024香蕉在线观看| 国产日韩一区二区三区精品不卡| 宅男免费午夜| 久久热在线av| a级片在线免费高清观看视频| 自拍欧美九色日韩亚洲蝌蚪91| 成年av动漫网址| 欧美精品一区二区免费开放| 另类精品久久| 成人手机av| 国产亚洲av高清不卡| av网站免费在线观看视频| 欧美 亚洲 国产 日韩一| 亚洲av电影在线进入| 亚洲精品久久午夜乱码| 一本一本久久a久久精品综合妖精| 一级毛片电影观看| 国产欧美日韩一区二区三 | 国产99久久九九免费精品| 国产不卡av网站在线观看| 国产精品欧美亚洲77777| 青春草视频在线免费观看| 久久久久久亚洲精品国产蜜桃av| 91成人精品电影| 黄色片一级片一级黄色片| 飞空精品影院首页| 99热国产这里只有精品6| 欧美亚洲日本最大视频资源| 18在线观看网站| 午夜激情av网站| cao死你这个sao货| 男人爽女人下面视频在线观看| 欧美日韩亚洲综合一区二区三区_| 免费一级毛片在线播放高清视频 | 久久亚洲国产成人精品v| av一本久久久久| 成人午夜精彩视频在线观看| 国产视频首页在线观看| 免费在线观看完整版高清| 午夜久久久在线观看| 亚洲第一av免费看| 日韩av免费高清视频| 国产真人三级小视频在线观看| 中文字幕人妻熟女乱码| 超碰成人久久| 国产伦理片在线播放av一区| 视频区图区小说| 久久精品亚洲熟妇少妇任你| 亚洲第一青青草原| 亚洲欧美成人综合另类久久久| 亚洲欧美日韩高清在线视频 | 国产熟女午夜一区二区三区| 亚洲 欧美一区二区三区| 亚洲成人国产一区在线观看 | 国产精品一区二区在线观看99| av国产精品久久久久影院| 日本黄色日本黄色录像| 国产av国产精品国产| 亚洲成av片中文字幕在线观看| 欧美日韩视频高清一区二区三区二| 亚洲男人天堂网一区| 国产一区二区 视频在线| 久久天堂一区二区三区四区| 久久综合国产亚洲精品| 黄色视频不卡| 日日摸夜夜添夜夜爱| 国产三级黄色录像| 91九色精品人成在线观看| 久久女婷五月综合色啪小说| 亚洲欧美清纯卡通| 人成视频在线观看免费观看| 97人妻天天添夜夜摸| 永久免费av网站大全| 曰老女人黄片| 十八禁人妻一区二区| 一区二区三区激情视频| 美女主播在线视频| 午夜免费成人在线视频| 精品欧美一区二区三区在线| 久热这里只有精品99| 精品一区在线观看国产| 中文字幕人妻丝袜一区二区| 少妇猛男粗大的猛烈进出视频| 国产免费现黄频在线看| 99久久综合免费| 90打野战视频偷拍视频| 中文字幕精品免费在线观看视频| 大话2 男鬼变身卡| 少妇 在线观看| 一级,二级,三级黄色视频| 欧美成人午夜精品| 观看av在线不卡| 好男人电影高清在线观看| 菩萨蛮人人尽说江南好唐韦庄| 男女午夜视频在线观看| 久久久国产一区二区| 久久久精品免费免费高清| 欧美大码av| 男女床上黄色一级片免费看| 午夜福利视频精品| 久久中文字幕一级| 丝袜脚勾引网站| 国产黄色免费在线视频| 超碰97精品在线观看| 欧美日韩一级在线毛片| 亚洲精品一区蜜桃| 国产一区亚洲一区在线观看| 国产主播在线观看一区二区 | 午夜福利在线免费观看网站| 日本欧美国产在线视频| 国产精品 国内视频| 国产精品国产三级国产专区5o| 亚洲五月婷婷丁香| 好男人电影高清在线观看| 桃花免费在线播放| 午夜福利乱码中文字幕| 91老司机精品| 亚洲人成77777在线视频| 中文字幕精品免费在线观看视频| 亚洲国产欧美一区二区综合| 婷婷色av中文字幕| 黄色片一级片一级黄色片| 男女午夜视频在线观看| 亚洲欧洲日产国产| 99re6热这里在线精品视频| 国产1区2区3区精品| 欧美日韩亚洲综合一区二区三区_| 欧美精品啪啪一区二区三区 | 18禁黄网站禁片午夜丰满| 精品亚洲成a人片在线观看| 啦啦啦在线观看免费高清www| 午夜福利一区二区在线看| 久久人人爽av亚洲精品天堂| 90打野战视频偷拍视频| 亚洲精品日本国产第一区| 免费看十八禁软件| 晚上一个人看的免费电影| 色婷婷av一区二区三区视频| 久久亚洲精品不卡| 久久99热这里只频精品6学生| 波野结衣二区三区在线| 曰老女人黄片| 男人操女人黄网站| 欧美中文综合在线视频| 18禁国产床啪视频网站| av国产精品久久久久影院| 天天添夜夜摸| 亚洲欧美一区二区三区国产| 婷婷色综合大香蕉| 精品亚洲成a人片在线观看| 国产男女内射视频| 人人妻人人爽人人添夜夜欢视频| 王馨瑶露胸无遮挡在线观看| 久久99一区二区三区| 国产视频一区二区在线看| 精品熟女少妇八av免费久了| 日本91视频免费播放| 十分钟在线观看高清视频www| 亚洲国产看品久久| 精品一区二区三区av网在线观看 | 男女午夜视频在线观看| 欧美国产精品一级二级三级| 黄色视频不卡| 又黄又粗又硬又大视频| 免费高清在线观看视频在线观看| 亚洲精品国产av蜜桃| 亚洲成国产人片在线观看| 国产主播在线观看一区二区 | 亚洲九九香蕉| 又紧又爽又黄一区二区| 视频区图区小说| 久久国产精品男人的天堂亚洲| 国产精品一区二区在线不卡| 国产亚洲欧美在线一区二区| 亚洲一区中文字幕在线| 一区在线观看完整版| 日本91视频免费播放| 久久久久精品国产欧美久久久 | 免费观看av网站的网址| 黄网站色视频无遮挡免费观看| 人人妻人人澡人人看| 欧美精品高潮呻吟av久久| 国产高清国产精品国产三级| 夫妻性生交免费视频一级片| 狂野欧美激情性bbbbbb| 18禁黄网站禁片午夜丰满| 欧美在线黄色| 久久综合国产亚洲精品| 亚洲九九香蕉| 日韩中文字幕欧美一区二区 | 80岁老熟妇乱子伦牲交| 国产免费视频播放在线视频| 交换朋友夫妻互换小说| 国产一级毛片在线| 亚洲国产欧美在线一区| 亚洲男人天堂网一区| 丰满少妇做爰视频| 国产成人影院久久av| 成年av动漫网址| 高清av免费在线| 亚洲精品美女久久av网站| 老司机深夜福利视频在线观看 | 日本av手机在线免费观看| 成在线人永久免费视频| 国产精品免费视频内射| 午夜老司机福利片| av欧美777| 一区二区三区四区激情视频| 日本a在线网址| 免费看av在线观看网站| 成年女人毛片免费观看观看9 | 国产精品三级大全| 成年人免费黄色播放视频| 五月天丁香电影| 美女大奶头黄色视频| 高清欧美精品videossex| 国产黄色免费在线视频| 亚洲少妇的诱惑av| 国产片内射在线| 人人妻人人澡人人爽人人夜夜| 国产精品一区二区精品视频观看| 亚洲欧洲国产日韩| 国产精品久久久人人做人人爽| 国产成人a∨麻豆精品| 母亲3免费完整高清在线观看| 午夜精品国产一区二区电影| 国产亚洲午夜精品一区二区久久| 精品少妇黑人巨大在线播放| 五月天丁香电影| 99香蕉大伊视频| 亚洲五月色婷婷综合| 性色av乱码一区二区三区2| 国产视频首页在线观看| 午夜久久久在线观看| 国产无遮挡羞羞视频在线观看| 久久精品久久久久久噜噜老黄| 国产亚洲av片在线观看秒播厂| www.精华液| 在线精品无人区一区二区三| 男男h啪啪无遮挡| 国产有黄有色有爽视频| 午夜福利影视在线免费观看| 夫妻性生交免费视频一级片| 啦啦啦在线免费观看视频4| 日韩熟女老妇一区二区性免费视频| 美女福利国产在线| 99国产综合亚洲精品| 我要看黄色一级片免费的| 99国产精品一区二区蜜桃av | 自线自在国产av| 欧美黄色淫秽网站| 国产又爽黄色视频| 一级毛片电影观看| 99re6热这里在线精品视频| 国产1区2区3区精品| 美女脱内裤让男人舔精品视频| 国产片特级美女逼逼视频| 18在线观看网站| 中文字幕制服av| 99久久精品国产亚洲精品| 欧美精品高潮呻吟av久久| 黑人猛操日本美女一级片| 中文字幕人妻丝袜一区二区| 99国产精品99久久久久| 性色av乱码一区二区三区2| 水蜜桃什么品种好| 国产色视频综合| 日韩制服丝袜自拍偷拍| 久久久久久亚洲精品国产蜜桃av| 男人添女人高潮全过程视频| 男女国产视频网站| 久久鲁丝午夜福利片| 欧美乱码精品一区二区三区| 成年美女黄网站色视频大全免费| 丝袜脚勾引网站| 久久99精品国语久久久| 中文字幕亚洲精品专区| 久久久久视频综合| 午夜免费男女啪啪视频观看| 啦啦啦中文免费视频观看日本| 国产av一区二区精品久久| 麻豆av在线久日| 一边摸一边抽搐一进一出视频| 亚洲黑人精品在线| 久久久久国产精品人妻一区二区| 69精品国产乱码久久久| 极品少妇高潮喷水抽搐| 午夜久久久在线观看| www.999成人在线观看| 国产av国产精品国产| 香蕉丝袜av| 97精品久久久久久久久久精品| 婷婷色综合大香蕉| 在线观看一区二区三区激情| √禁漫天堂资源中文www| 亚洲国产精品999| 亚洲色图综合在线观看| 亚洲欧美精品自产自拍| 免费看av在线观看网站| 一本—道久久a久久精品蜜桃钙片| 免费在线观看完整版高清| av国产久精品久网站免费入址| 久久精品成人免费网站| 久久精品国产a三级三级三级| www.熟女人妻精品国产| 亚洲av日韩精品久久久久久密 |