• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Classification of epilepsy using computational intelligence techniques

    2016-03-20 07:11:17KhurrmQziLmBoXioGoxingOuyngXunheYin

    Khurrm I.Qzi*,H.K.LmBo XioGoxing Ouyng,Xunhe Yin

    aDepartment of Informatics,King's College London,Strand,London,WC2R 2LS,United Kingdom

    bState Key Laboratory of Cognitive Neuroscience and Learning,School of Brain and Cognitive Sciences,Beijing Normal University,No.19,XinJieKoWai St., HaiDian District,Beijing,100875,PR China

    cSchool of Electronic and Information Engineering,Beijing Jiaotong University,Beijing,PR China

    Available online 13 October 2016

    Classification of epilepsy using computational intelligence techniques

    Khurram I.Qazia,*,H.K.Lama,Bo Xiaoa,Gaoxiang Ouyangb,Xunhe Yinc

    aDepartment of Informatics,King's College London,Strand,London,WC2R 2LS,United Kingdom

    bState Key Laboratory of Cognitive Neuroscience and Learning,School of Brain and Cognitive Sciences,Beijing Normal University,No.19,XinJieKoWai St., HaiDian District,Beijing,100875,PR China

    cSchool of Electronic and Information Engineering,Beijing Jiaotong University,Beijing,PR China

    Available online 13 October 2016

    This paper deals with a real-life application of epilepsy classification,where three phases of absence seizure,namely pre-seizure,seizure and seizure-free,are classified using real clinical data.Artificial neural network(ANN)and support vector machines(SVMs)combined with supervised learning algorithms,and k-means clustering(k-MC)combined with unsupervised techniques are employed to classify the three seizure phases.Different techniques to combine binary SVMs,namely One Vs One(OvO),One Vs All(OvA)and Binary Decision Tree(BDT),are employed for multiclass classification.Comparisons are performed with two traditional classification methods,namely,k-Nearest Neighbour(k-NN)and Naive Bayes classifier.It is concluded that SVM-based classifiers outperform the traditional ones in terms of recognition accuracy and robustness property when the original clinical data is distorted with noise.Furthermore,SVM-based classifier with OvO provides the highest recognition accuracy,whereas ANN-based classifier overtakes by demonstrating maximum accuracy in the presence of noise.

    Absence seizure;Discrete wavelet transform;Epilepsy classification;Feature extraction;k-means clustering;k-nearest neighbours;Naive Bayes;Neural networks;Support vector machines

    1.Introduction

    Epilepsy is a neurological condition such that it affects brain and the nervous system.It is a very commonly known neurological disorder and approximately 1%of general population is affected[1].Only in the UK,around 1 in 100,more than half a million people suffer from epilepsy.There can be many causes of epilepsy and sometimes it is not possible to identify them.In the domain of epilepsy,seizure is referred to as an epileptic seizure and brain is the source.During an epileptic seizure normal functioning of the brain is disturbed for that certain time period,causing disruption on signalling mechanism between brain and other parts of the body.Theseseizures can put epilepsy patients at higher risk for injuries including fractures,falls,burns and submersion injuries,which are very common in children[2].These injuries happen because seizure can happen anytime and anywhere without prior warning and the sufferer would continue his or her activity with an unconscious mind.If a system can effectively predict the pre-seizure phase(the transition time of the brain towards developing seizure),it could then generate an early warning alarm so that precautions can be taken by the sufferer.

    Absence seizure is one from many forms of generalized epileptic seizures in which larger part of the brain is disturbed. These seizures are very short and sometime may go un-notice. The patient seems confused and may not remember the seizure afterwards.The complex spike-and-wave patterns generated by the brain during these seizures can be recorded on the electroencephalogram(EEG)and a neurologist can identify the three absence seizure phases namely seizure-free,preseizure and seizure[3,4].To automate this process,EEG data is converted into a digital format and fed into a computerizedseizure detection(classification)system,which can automatically recognize the input pattern.The two core modules of this classification system are:feature extraction and design of a classifier using these features.A feature extraction method extracts the most discriminative information from the EEG recordings,which means an ideal feature can have the property of differentiating among three phases of absence seizure.

    In literature,there is a long list of methods that can be used to extract features from the EEG signals.These methods include Fourier Transforms(FT);good for analysing stationary signals,and Time Frequency Distribution(TFD);provides a tool for examining continuous segments of EEG signals. However,EEG signals are non-stationary in nature and conventional frequency analysis methods may not capture the full details of the brain signals.Lyapunov exponents;discussed the detection and prediction of epileptic seizure in Refs.[5], analysis of correlation structure[6],and high order spectral analysis(HOS)[7]are the examples of non-linear methods for EEG signal analysis in the domain of epilepsy.Advances in wavelet theory has also made it a very suitable for bio-medical signal processing.It has a built-in advantage of capturing repeated and irregular patterns.It can also deal with the analysis of transient and sudden signal changes[8,9].This is possible because this technique provides variable window size, narrow at high and wide at low frequency levels.Furthermore [8],has discussed several different methods for EEG signal analysis and concluded that Wavelet Transform(WT)has more advantages over other methods.The next step in designing the classification system is to combine these extracted features with an appropriate learning method to design a classifier.These methods can be divided into traditional classification methods and modern learning algorithms also known as computationally intelligent algorithms or machine learning algorithms.Bayesian methods based on statistical theory,k-nearest neighbour(k-NN)and decision trees based on logical branching,are considered to be in the category of traditional classifiers.On the other hand,Support Vector Machine(SVM),Artificial Neural Network(ANN),k-Means Clustering(k-MC)and Self-organising Maps(SOMs) exhibit intelligent behaviour by learning the complex and nonlinear patterns hidden inside the input feature vector,are considered to be computationally intelligent approaches. Below is the brief introduction about all the methods used in this research.

    Classification systems based on Bayesian statistical theory have been widely and successfully used commercially[10]. This method gives a way to represent sensory evidence;features extracted from the raw data,and prior information about the problem in hand that is collected from domain knowledge. Considering the equal prior probabilities for all classes and hence,ignoring the hassle of obtaining the domain knowledge, the analysis becomes very straight forward.Although,it is a powerful and simple rule to handle and implement,yet estimating posterior probabilities from the data is a non-trivial task[11]and the distribution of data may not be uniform.k-NN rule as the name implies,classify an observation by giving it a label after probing the labels on all the k-NNs and making decision based on majority voting.Usually,Euclidean distance is used to measure the distances between the neighbouring instances.This algorithm mostly provides an acceptable performance in many applications[10]such as visual category recognition[12]and is also very easy to implement.However, k-NN algorithm suffers due to large memory requirements and also there is no logical way to choose the best k value,which would affect the classification problem and may not yield very good results.

    Inspired by the human brain functioning and architecture, McCulloch and Pitts in 1940s presented a logical threshold unit(LTU)[13].This basic idea of LTU has been generalized in many ways since then and it is the building block for modern ANNs.In 1962 a single layer perceptron neural network was introduced by Rosenblatt[14]by extending the idea of LTU along with a trainable network with adaptive elements[15].Despite the huge success of this neural network model,it was only limited to solve linearly separable problems.A traditional feed-forward neural network[13]has three types of layers,i.e.,input,hidden and output layers.Each layer consists of nodes connected in a layer-to-layer manner.The feed-forward neural networks can learn any smooth non-linear functions in a compact domain to any degree of accuracy and are considered to be universal approximators.They have many applications such as classification,prediction,clustering and approximation[13].The drawback of this technique is that it requires a lot of parameters to be tuned for training the neural network.Also,there is no set of rules for finding the number of hidden layers and neurons in these layers.Furthermore, ordinary neural networks suffer from the"over-fitting"and "local optimization"problems[16].

    The SVM method maps the non-linear and inseparable data from an input space into a higher dimensional feature space where the data would then be linearly separable[17].This task is accomplished by utilising the concept of separating hyper planes[17].Instead of computing a mapping function,the use of kernel function saves the computational demand especially for feature mapping function of higher dimensional space.The SVM algorithm aims to maximize the margin(the region separating the support vectors on either side of the hyperplane) and tries to find an optimal hyperplane.Hence,also called the maximal margin classifier.Although,the training parameters for SVM technique are very few,it can still be a computationally time consuming and highly complex.Nonetheless, SVM has a good generalization ability,solution to the overfi tting problem and also performs well in a high dimensional feature space.

    k-MC is an unsupervized learning technique that clusters samples based on similar properties.Whereas,k represents the number of clusters or classes.Within a cluster the samples are similar but different from the samples grouped in other clusters.This is an iterative process in which samples are grouped together according to their closet mean and cluster positions are adjusted until these positions do not change for some iterations.The advantage of this technique is that it does not require labels with the input examples.The convergence is also faster if k is small.However,if clusters are of differentsizes,densities and not globular then the results are poor.The results may also vary depending on the initial location of the cluster centres.

    From the above discussion,we can conclude that the advantage of using machine learning approaches over the traditional classification methods is that the classification system does not need to know much about the input data in the beginning but the characteristics of the input data are learnt by the learning algorithm.Once the classification system has been built,replacing a learning algorithm with any other would not require any changes in the system.Considering this advantageous factor,classifiers such as ANN and SVM have been widely used in designing the epilepsy detection systems [3,9,6,18].Also,features that are highly discriminative not only yields better recognition accuracy but also speeds up the process.Recently[3],has shown a very good recognition accuracy of seizure phases by extracting features in the time domain as well as in the frequency domain using FT.

    In this paper,we first present a different strategy for extracting features from the EEG signals.In the time domain local features,based on windowing filtered mechanism,and global features,based on the entire signal length,are calculated.The signal is then decomposed into different levels using Discrete Wavelet Transform(DWT)and features at each level are calculated.In the next phase,we compare different machine learning algorithms in terms of their recognition accuracy based on the extracted features.Traditional classification methods provide an acceptable performance in many applications and can be used as a benchmark for other techniques [10].Comparisons have also been performed between both types of classification methods,traditional and intelligent algorithms.The robustness property has also been tested using noise-contaminated data.

    The organisation of this paper is as follows.Section 2 explains the mechanism of digital data acquisition from epilepsy patients and also discusses the process of feature extraction method for designing the epilepsy classification system.It then explains how the three computational intelligence techniques; SVM,ANN and k-MC work.Section 3 presents classification results that include design of classifiers using SVM,ANN and k-MC and evaluation of these classifiers against two traditional classification methods,NBC and k-NN.The paper ends with concluding remarks in Section 4.

    2.Epilepsy classification system

    Fig.1 shows the design cycle of epilepsy classification system used in this research.This is an iterative process in which the system is updated based on the output produced by the learning algorithms.The details of this design scheme are discussed in the following sections.

    2.1.Data acquisition

    Using 10-20 international standard electrode placement system,with all 19 electrodes(FP1,FP2,F3,F4,C3,C4 P3, P4,O1,O2,F7,F8,T3,T4,T5,T6,Fz,Cz,and Pz)the data has been collected from the epilepsy patients suffering from the absence seizure.These recordings were taken in the Peking University People's Hospital,China,and the patients(6 males and 4 females),aged 8-21 years old,have agreed to use this data for research and publishing results[3].To record the neural activity,EEG data was samples at 256 Hz,filtered from 0.5 to 35 Hz bands using Neurofile NT digital video EEG system[3].Extracted data has been divided into three phases of absence seizure;seizure-free,pre-seizure and seizure,by an epilepsy neurologist.These phases are separated based on the criteria of 1)the interval between the seizure-free phase and beginning point of seizures is greater than 15 s,2)the interval is 0 and 2 s prior to seizure onset,and 3)the interval is the first 2 s of the absence seizure[3].Each dataset has 110 samples and each sample size is 19X512.Fig.2 shows the example of 19-channel EEG recordings of these three seizure phases. From the figure,seizure-free and seizure phases are quite obvious and easier to classify.However,the brain shifts from seizure-free to an absence seizure(pre-seizure phase)looks almost the same and is hard to distinguish.The generalized spike-wave discharges with a repetition rate of 3 Hz are typically associated with clinical absence seizures.More details about the data collection can be located in Refs.[19-21].

    2.2.Feature extraction

    The purpose of this stage is to eliminate the redundancy in the EEG signals by selecting the discriminative features from the raw data.The process also helps in reducing the size of the input feature vector.In the case of EEG recordings,we have 19 channels from the 19 EEG electrodes(sensors)as shown in Fig.2.The research in absence epilepsy has demonstrated that not all 19 channels are of the same importance.Successful experiments in Ref.[3]has also established that only 10 electrodes(FP1,FP2,F3,F4,C3,C4,F7,F8,T3 and T4)out of total 19 have the properties that can help in classifying the three seizure phases.A feature vector was then formed based on these 10 selected channels.

    To capture the brain functioning over a period of time,local and global values of the same features were calculated from the EEG signals in the time domain.Global values were calculated based on the entire signal length and local features were extracted using a windowing function.The importance of the features in the time domain is evident from Fig.2,in which there are no sudden or abrupt changes in the brain recordings during the seizure-free and pre-seizure phases.Such activities are very well captured by the time domain analysis.Energy (E),range(R),standard deviation(SD),sum of absolute values (SAV),mean absolute values(MAV)and variance(Var)of each selected channel were extracted in the time domain to form the part of the feature vector.

    Fig.1.Design cycle of epilepsy classification system.

    Fig.2.EEG recordings.Seizure-Free(A),Pre-Seizure(B)and Seizure(C)phases.

    DWT decomposes the signal into several levels and each level represents a particular coarseness of the signal.At each level,the signal is passed through the high pass filter(HPF), which acts as the mother wavelet,and the low pass filter(LPF) that acts as a mirror version of the corresponding HPF.The output of each level is the downsized signal by a factor of 2. Approximation(A1),consists of low frequency components, and detail(D1),consists of high frequency components,are the coefficients produced by LPF and HPF respectively.For the next level of decomposition,A1 is further decomposed and the whole process is repeated until the desired level of decomposition is achieved.For one dimensional signal x(t)the continuous wavelet transform(CWT)is defined by Morlet-Grossman[22]:

    where ψ is the analysing wavelet,a is the scalar parameter, and b is the position parameter.The CWT is converted into discrete wavelet by using binary scale factors.Hence,from(1) [23]:

    Which means A1 and D1 are the first split of signal x(t). Then from using A1 the next split is A2 and D2 and the process continues until the desired level is reached.Fig.3 depicts the DWT decomposition scheme of a signal of length n and↓2shows that the signal is downsized by a factor of 2 at each level.

    We have used Daubechies D4 wavelets,which is most widely used DWT for EEG signal analysis.D4 uses four scaling and wavelet coefficients[9].The scaling coefficients (LPF)are:

    Fig.3.DWT decomposition scheme.

    And the wavelet coefficients(HPF)are:

    EEG signal has been decomposed into 5-levels and at each level energy,range and standard deviation has been calculated. In addition to these features,waveform length(WL),average amplitude change(AAC)and difference absolute standard deviation(DASD)[24]have also been computed at each level of decomposition to form the rest of the feature vector. Combining all the features from time domain as well as from the decomposed signal resulted in a large feature vector,which is considered to be computationally expensive in the classi fication process and can also generate poor recognition accuracy.Fig.4 depicts the whole process of feature extraction discussed above.As from Section 2.1,the individual electrode provides 512 brain readings(values)and only 10 channels are selected to build the feature vector.Hence,form each electrode,192 local values using a window size of 16 and 6 global feature values are extracted.The values of individual electrode is then decomposed to 5-levels.Dj…DLrepresent the detail values and ALrepresents approximate values as shown in Fig.3.Where j=1…L and L=5.Features are extracted at each level giving 36 values per electrode.Principal component analysis(PCA)has been used to reduce the dimension(d)of the feature vector.A 40 dimensional feature vector is decided to be used as an input to design the classifier and to perform the classification experiments.

    2.3.Computational intelligence techniques

    Intelligent techniques take feature vector as an input and use a learning strategy these algorithms learn the patterns in the features vector.After the learning(training)process hasfinished any new data sample is classified based on the trained classifier.This phase is called classification phase. According to the learning procedure,these techniques can be divided into supervised and unsupervised learning.In supervised learning,thealgorithm takesexampleinputs (training dataset)and their corresponding output(class labels).It then learns a general rule of mapping inputs to outputs.SVM and ANN follow this strategy.In an unsupervised learning mechanism,no output is attached with the input examples.The data is grouped and clustered based on their natural or similar characteristics.k-MC is an unsupervised learning algorithm.In this research,we have explored SVM,ANN and k-MC.

    2.3.1.Support vector machine

    A binary class classification problem y(i)∈[+1,-1]with a training set,and the corresponding outputs (class labels)can be stated as:

    WhereW∈?dis the weight vector and w0is the bias.However,not all the classification problems are linearly separable. Such problems are solved using the soft-margin method and the classifier is known as soft-margin classifier.Maximising the margin and minimising the number of misclassified points, we can formulate the constraint optimization problem for the soft-margin method as:

    Fig.4.Feature extraction process.

    Where ζ is a slack variable,which represents the upper bound training error and C>0 trades off margin size and training error.Replacing C with 0 in(5)becomes the linearly separable case.Non-linear Support Vector Classifiers(SVCs)use the kernel trick,which maps the original feature space to another higher dimensional space and through this mapping process the classification task may become easier.In this study we use the following three kernel functions[25].

    where K(.,.)represents a kernel function,d is the dimension of the feature vector,q is the degree of the polynomial and γ sets the influence boundary for the radial basis kernel function.SVM is a binary classifier that can only handle two-class classification problems.To overcome this shortcoming different strategies are used to combine the binary SVM classifiers to tackle multiclass classification problems.As in our case,we are required to classify three phases of absence seizure.We have compared three different methods to combine binary SVM classifiers,which are Binary Decision Tree(BDT),One Vs One(OvO),and One Vs All(OvA).

    BDT method requires C-1 classifiers for C-classes. Converting a three-class classification problem into a twoclass classification problem,SVM1 is trained on samples from seizure phase(class 3)with labels,-1,and samples from seizure-free and pre-seizure phases(classes 1&2)combined together as one class with labels,+1.SVM2 is then trained on samples from seizure-free phase(class 1)with labels,+1,and pre-seizure phase(class 2)samples with labels,-1.A test sampleX(*)is processed through both the SVMs and to obtain the category of this new sample the results of both SVMs are then combined as shown in Fig.5.The classification accuracy depends on the SVMs in the upper levels.binary SVM classifiers are required for C-classes using OvO method.Thus,for three-class problem,it requires 3 binary SVMs.SVM1 is trained on samples from class 1 (seizure-free)and class 2(pre-seizure),SVM2 on samples from class 2 and class 3(seizure),and SVM3 on samples from class 3 and class 1.For a new observation(sample)X(*),it chooses final class label with majority votes.This whole process is summarized in Fig.6.

    Fig.5.Three-class SVM-BDT.

    C-binary SVM classifiers are needed for C-class problem when using OvA method.Therefore,3 binary classifiers for the problem in hand.Three-class problem is again divided into two-class classification problem and the training procedure is shown in Fig.7,where class 1 is seizure-free,class 2 is preseizure and class 3 is seizure phase.A new sampleX(*)is passed through all three SVMs and the SVM having highest value will decide the final class label.

    2.3.2.Artificial Neural Networks

    A fully-connected three-layer feed-forward neural network (FFNN)is shown in Fig.8.Each layer is connected to its previous and the next layer.These connections are associated with connection weights.Hidden layers also contain an array of artificial neurons,called processing units and an activation function,f(.),for each neuron.A general feed-forward operation for classification purposes can be written as[16]:

    where the subscript j indexes units in the input layer and p indexes units in the hidden layer.wpjdenotes the input-tohidden layer weights at the hidden unit p.The subscript k indexes in the output layer and nHdenotes the number of hidden units.d represents the dimension of the ith input sample.The advantage of using FFNN compared with SVM is that for C-classes(outputs),the network can learn C discriminant functions zkand a training inputX(i)is assigned to the functionthat has the maximum value from all other functions.Usually,these functions are learnt through a wellknown backpropagation algorithm.The goal of this algorithm is to find a set of weight values for all the connections, such that it can minimize the error between the actual and desired output.A new sampleX(*)is classified using the trained FFNN.

    To achieve the maximum recognition accuracy of the three seizure phases,we have tried different number of hidden layers,hidden nodes and various combination of activation functions in this study.The best recognition accuracy was achieved by using 4 hidden-layer FFNN.The first hidden layer has 22 neurons and uses radial basis activation function.The second hidden layer has an array of 8 neurons and each node uses tangent sigmoid activation function.The third and fourth hidden layers contain 10 and 4 neurons respectively,and bothlayers are using logarithmic tangent sigmoid transfer function. The output layer is consisting of three neurons.Both input and output layers are using linear activation function.The FFNN takes the samples of size d as an input along with the corresponding class labels and then produces the output of size 3. The class labelling scheme has been shown in Table 1.The training parameters and procedure for this 4 hidden-layer FFNN architecture have been provided in the results section.

    Fig.6.Three-class SVM-OvO.

    Fig.7.Three-class SVM-OvA.

    Fig.8.Three-layer feed-forward fully connected neural network.

    2.3.3.k-means clustering

    The problem of clustering can be defined as the sum of square Euclidean distances of each cluster from its meanmk:

    Table 1Class labelling scheme for FFNN.

    This criterion defines clusters as their mean vectorsmkin the sense that it minimizes the sum of the squared lengths of the error,X(i)-mk.Where C represents the total number of clusters(classes),X(i)is the ith sample,and Dkis the kth cluster.Hence,a partition is optimal if it minimizes Jeand also called minimum variance partition.In our case,the data was clustered into three groups(seizure-free,pre-seizure and seizure phases)and a new sample was assigned to the cluster that has minimum distance from these three clusters.

    It is worth mentioning that after the training process,all the parameters and weights of the classifiers are fixed values and can be used in the real-life applications.The computational cost for the classifiers to give out the results are all very low and can be ignored in most of the cases.

    3.Classification results

    From Section 2.1,each dataset(seizure-free,pre-seizure and seizure)has 110 samples.To design classifiers using extracted features from the EEG signals as discussed inSection 2.2,and to perform the classification experiments, each dataset is divided into 22 testing samples and 88 training samples.The training dataset is used to train the classifier. After the training process,the classifier is tested using the testing dataset to evaluate its recognition accuracy for any new pattern(sample).The percentage of recognition accuracy for both training and testing datasets are calculated separately based on correctly classified samples within each individual dataset.

    3.1.Classifiers

    To perform the experiments with SVM-based classifiers,a soft-margin classifier has been used in all the multiclass SVM classifiers.When polynomial kernel function(7)is used,we choose q=5.When radial basis kernel function(8)is used, we choose γ=0.2.These values are used in all experiments which are achieved by trial and error approach for the best performance.

    The FFNN architecture discussed in Section 2.3.2 was trained with training samples along with their corresponding class labels for 30 epochs and with initial learning rate of 10-4using Levenberg-Marquardt backpropagation algorithm. Mean squared error criteria was used to calculate the error between the outputs produced by the network and the actual class labels.For each successful step,the learning rate was decreased by 10-1and for every failure,it was incremented by 5.The network takes 40 dimensional feature vector as an input,and produces the output of size 3.(1,0,0)represents an observation from class seizure-free,(0,1,0)corresponds to a sample from pre-seizure phase and(0,0,1)classifies a seizure phase.

    For k-MC,100 iterations were performed to determine the cluster centres.However,clustering is a random process, which starts with random cluster centres.During these iterations,if the clusters did not change their positions,new clusters centres were determined to avoid the local minima problem.This process replicated itself 5 times.

    3.2.Evaluation

    To evaluate the success of our feature extraction scheme and to equate the recognition accuracybetween learning algorithms and traditional classification methods,we have also selected to perform experiments with naive Bayes and k-nn classification methods.To estimate the data distribution for naive Bayes classifier,we have employed two ways to estimate the input data distribution.The one uses the normal Gaussian function (Naive Bayes with Gaussian function NBN)and the other uses the kernel smoothing function(Naive Bayes with kernel smoothing function NBK).k-NN has been tested for four different values of k(i.e.,1,3,5 and 7).Class labels 1,2,and 3 have been used for seizure-free,pre-seizure and seizure samplesrespectively,forboththetraditionalclassificationmethods.

    The recognition accuracy(classification performance)of all the methods are evaluated.Firstly,the experiments were performed using the original dataset,without any noise.In the EEG process the noise however is inevitable due to certain uncertainties in the measurement process,such as,signal conversion,filtering,amplifying and environment conditions. To reflect these factors present in the real-life patient's clinical data,how the underlying seizure recognition system would behave in the presence of noise and also to investigate the robustness property of each algorithm,the data has been contaminated with different levels(0.05,0.1,0.2,0.5 and 1.0) of random noise.

    The recognition accuracy of all the methods are presented in Tables 2-7,where SVM-OvA represents SVM classifier with one vs all strategy,SVM-OvO is SVM classifier with one vs one method and SVM-BDT shows the results obtained using SVM classifier with binary decision tree method.Linear, polynomial and radial basis kernel functions show the results against each individual kernel method discussed in Section 2.3.1.NN is the neural network classifier consisting of four hidden layer architecture,NBC is naive Bayes classification method,which uses two different density estimation methods, normal(NBN)and kernel smoothing function(NBK).k-NN is shown with different values of k(i.e.,1,3,5 and 7).Andfinally,k-MC shows the k-means clustering results.

    In Table 2,the recognition accuracy(%)of both training and testing datasets is shown where"Average"represents the average accuracy for all three classes(seizure phases).It can be seen that the highest average testing accuracy is 93.9394% which is achieved by SVM-OvO using polynomial kernel function and the corresponding average training accuracy is 99.6970%.Whereas,1-NN has obtained the best average training recognition accuracy of 100%.On the other hand,the lowest average testing accuracy is 66.6667%which is achieved by k-MC.Talking about successfully predicting the preseizure phase,SVM-OvO with polynomial kernel function is again leading with 90.9091%testing recognition accuracy. Also,traditional classification methods are not able to predict the pre-seizure phase with good recognition accuracy.NBC with kernel function has improved the average testing accuracy from 69.1358%to 72.8395%.However,NBC provides a poor performance in predicting pre-seizure phase.

    The fact that the real-life data will not be smooth and it will contain noise due to certain factors mentioned above.Tables 3-7 summarizes the results of testing data classification when original testing dataset is contaminated with different levels of noise.Addition of noise is a random process.Thus to collect the statistical information,each experiment was repeated 10-times on the test dataset only.These experiments were performed without training the classifier on the noisy dataset."Worst","average","best"and"std"(standard deviation)show overall values for all the three seizure phases.The worst individual is the lowest individual accuracy obtained after these experiments.

    Generally,the recognition accuracy is decreasing when the noise level is increasing.The recognition accuracy of SVM with polynomial kernel function has badly degraded in all the multiclass SVM methods.However,SVM with linear kernel function has shown good robustness property in the presence of different noise levels.Among all the classification methods,

    the NN classifier has shown the maximum recognition accuracy of 85.8943%at highest noise level of 1.From the discussion and the results presented,it is fair to conclude that NN has a good generalization ability which is not very much degraded under the increasing levels of noise.

    Table 2Summary of training and testing recognition accuracy for EEG signals with original datasets.

    Table 3Summary of testing recognition accuracy for EEG signals under testing dataset subject to noise level of 0.05.

    Table 4Summary of testing recognition accuracy for EEG signals under testing dataset subject to noise level of 0.1.

    Table 5Summary of testing recognition accuracy for EEG signals under testing dataset subject to noise level of 0.2.

    Table 6Summary of testing recognition accuracy for EEG signals under testing dataset subject to noise level of 0.5.

    Table 7Summary of testing recognition accuracy for EEG signals under testing dataset subject to noise level of 1.

    4.Conclusion

    This paper has presented different supervised(SVM and NN)and unsupervised(k-MC)learning algorithms for classifying epilepsy seizure phases.Computationally intelligent techniques NN and SVM have proved to be very good in recognising and classifying the complex and complicated patterns in the input data(EEG signals).The performance of these algorithms has been compared with two traditional classification methods,NBC and k-NN.It can be concluded that NN and SVM have demonstrated the best recognition accuracy compared with traditional classification methods and an unsupervised learning algorithm,k-MC.Furthermore,NN and SVM both showed robustness property by maintaining best results even when the input data is contaminated with different noise levels.In addition,the feature extraction method is also introduced in the paper,which is able to gain rich information of the signal at a moderate dimension of the feature vector.Also,the NN method outperforms the SVM methods for some of the noisy dataset cases.In future,more classification methods based on neural networks such as selforganising maps and deep learning architecture will be explored and try to find the method,which is best at both noisy case and noise free case.

    Acknowledgment

    The work described in this paper was partly supported by King's College London,China Scholar Council,National Natural Science Foundation of China(No.61172022)and Foreign Experts Scheme of China(No.GDW20151100010).

    [1]P.Kwan,T.J.Brodie,Neurother 6(2006)397-406.

    [2]E.C.Wirrell,Epilepsia 47(1)(2006)79-86.

    [3]H.K.Lam,Udeme Ekong,Bo Xiao,Gaoxiang Ouyang,Hongbin Liu, K.Y.Chan,Sai Ho Ling,Neurocomputing 149(Part C)(2015) 1177-1187.

    [4]U.Ekong,H.K.Lam,B.Xiao,G.Ouyang,H.Liu,K.Y.Chan,S.H.Ling, Neurocomputing 199(2016)66-76.

    [5]S.Osowski,B.Swiderski,A.Cichocki,A.Rysz,COMPEL Int.J. Comput.Math.Electr.Electron.Eng.Emerald 26 (5)(2007) 1276-1287.

    [6]James R.Williamson,Daniel W.Bliss,and David W.Browne,Epileptic seizure prediction using the spatiotemporal correlation structure of intracranial EEG,IEEE International Conference on Acoustics,Speech and Signal Processing(ICASSP),(2011)pp.665-668.

    [7]K.C.Chua,V.Chandran,U.R.Acharya,C.M.Lim,J.Med.Syst.35(6) (2011)1563-1571.Springer.

    [8]Amjad S.Al-Fahoum,Ausilah A.Al-Fraihat,ISRN Neurosci.2014 (2014)7.Hindawi Publishing Corporation,Article ID 730218.

    [9]Ibrahim Omerhodzic, Samir Avdakovic, Amir Nuhanovic, Kemal Dizdarevic,Kresimir Rotim,Energy distribution of EEG signal components by wavelet transform,in:Dumitru Baleanu(Ed.),Wavelet transforms and their recent applications in Biology and Geoscience, 2012,p.310.InTech ISBN:978-953-51-0212-0.

    [10]Anil K.Jain,Robert P.W.Dulin,Jianchang Mao,IEEE Tran Pattern Anal Mach.Intell.22(1)(2000)4-37.

    [11]G.P.Zhang,IEEE Trans.Syst.Man Cybern.Soc.Appl.Rev.30(4) (2000)451-462.

    [12]H.Zhang,A.C.Berg,M.Maire,J.Malik,IEEE Comput.Soc.Conf. Comput.Vis.Pattern Recognit.2(2006)2126-2136.

    [13]Anil K.Jain,Jianchang Mao,K.M.Mohiuddin,IEEE Comput.Sci.Eng. 29(3)(1996)31-44.

    [14]G.L.Marcialis,F.Roli,Pattern Recognit.Lett.26(12)(2005) 1830-1839.

    [15]F.Ham,I.Kostanic,Principles of Neurocomputing for Science and Engineering,McGraw-Hill,New York,2001.

    [16]Richard O.Duda,Peter E.Hart,David G.Stork,Pattern Classification, second ed,John Wiley and Sons Inc,U.K,2006,p.680.ISBN:978-81-265-116-7.

    [17]S.Yaman,J.Pelecanos,W.Zhu,Unifying PLDA and polynomial kernel SVMS,in:Proceedings of the IEEE International Conference on Acoustics,Speech and Signal Processing(ICASSP),IEEE,2013, pp.7698-7701.

    [18]Abdulhamit Subasi,Ahmet Alkan,Etem Koklukaya,M.Kemal Kiymik, Neural Netw.18(7)(2005)985-997.

    [19]M.Thulasidas,C.Guan,J.Wu,IEEE Trans.Neural Syst.Rehabil.Eng. 14(1)(2006)24-29.

    [20]T.C.Ferree,P.Luu,G.S.Russell,D.M.Tucker,Clin.Neurophysiol.112 (3)(2001)536-544.

    [21]O'Keefe,M.L.Recce,Hippocampus 3(3)(1993)317-330.

    [22]S.Sanei,J.A.Chambers,EEG Signal Processing,John Wiley and Sons Ltd,2007,p.312.ISBN:978-0-470-02581-9.

    [23]X.Zhang,W.Diao,Z.Cheng,Wavelet Transform and Singular Value Decomposition of EEG Signal for Pattern Recognition of Complicated Hand Activities,Springer,2007,pp.294-303 vol.4561 of the series lecture notes in computer science.

    [24]Angkoon Phinyomark,Pornchai Phukpattaranont,Chusak Limsakul, Expert Syst.Appl.39(2012)7420-7431.

    [25]Gareth James,Daniela Witten,Trevor Hastie,Robert Tibshirani,An Introduction to Statistical Learning with Application in R,Springer Science and Business Media,New York,2013,p.440.ISBN:978-1-4614-7138-7.

    Bo Xiaoreceived bachelor and master(Hons.)degree from Chongqing University,China,in 2010 and 2013, respectively.He is currently working towards his Ph.D.degree in King's College London.His current research interests include computational intelligence, T-S fuzzy model based fuzzy control,polynomial fuzzy model based fuzzy control and interval type-2 fuzzy logic.

    H.K.Lamreceived the B.Eng.(Hons.)and Ph.D.degrees from the Department of Electronic and Information Engineering,The Hong Kong Polytechnic University,in 1995 and 2000,respectively.During the periodof2000and2005,heworkedwiththeDepartment of Electronic and Information Engineering at The Hong Kong Polytechnic University as Post-Doctoral Fellow and Research Fellow respectively.He joined as a LectureratKing'sCollegeLondonin2005andcurrently a Reader.His current research interests include intelligent control systems and computational intelligence.

    Xunhe Yinreceived the bachelor and doctoral degree from Harbin Science and Technology University and Harbin Institute of Technology,in 1989 and 2000, respectively.Currently,he works in Beijing Jiaotong University,Beijing,China as Professor.His current research interests include networked control systems; communication,control,and security in smart grid; controland security of cyber-physical system; communication and control technologies in smart traffic systems;intelligent control theory with applications for communication,networks and other areas. He has published more than 50 journals and conference papers.

    Gaoxiang Ouyangreceived the B.S.degree in automation and the M.S.degree in control theory and control engineering both from the Yanshan University, Hebei,China,in 2002 and 2004,respectively,and the Ph.D.degree from the Department of Manufacturing Engineering,City University of Hong Kong,in 2010. He currently serves as an Associate Professor in the School of Brain and Cognitive Sciences,Beijing Normal University,Beijing,China.His research interests include biosignal analysis,neural engineering, and dynamics system.

    *Corresponding author.

    E-mail addresses:khurram_ishtiaq.qazi@kcl.ac.uk(K.I.Qazi),hak-keung. lam@kcl.ac.uk(H.K.Lam),bo.xiao@kcl.ac.uk(B.Xiao),ouyang@bnu.edu. cn(G.Ouyang),xhyin@bjtu.edu.cn(X.Yin).

    Peer review under responsibility of Chongqing University of Technology.

    http://dx.doi.org/10.1016/j.trit.2016.08.001

    2468-2322/Copyright?2016,Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    Copyright?2016,Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    欧美老熟妇乱子伦牲交| 国产精品香港三级国产av潘金莲| 少妇的丰满在线观看| 久久人妻熟女aⅴ| 精品人妻1区二区| 夜夜骑夜夜射夜夜干| av片东京热男人的天堂| 国产免费现黄频在线看| 丝瓜视频免费看黄片| av免费在线观看网站| 最新美女视频免费是黄的| 最近最新免费中文字幕在线| 黄色视频不卡| 少妇 在线观看| 国精品久久久久久国模美| 如日韩欧美国产精品一区二区三区| 99国产综合亚洲精品| kizo精华| 亚洲全国av大片| 国产精品1区2区在线观看. | 999精品在线视频| 欧美精品av麻豆av| 人人妻人人爽人人添夜夜欢视频| 手机成人av网站| 9191精品国产免费久久| 色视频在线一区二区三区| 国产av又大| 国产三级黄色录像| 国产欧美日韩一区二区三区在线| 国产av国产精品国产| 99国产综合亚洲精品| 欧美+亚洲+日韩+国产| 亚洲自偷自拍图片 自拍| 久久午夜综合久久蜜桃| 亚洲成国产人片在线观看| 大香蕉久久成人网| 欧美精品一区二区大全| 久久人妻福利社区极品人妻图片| 91精品三级在线观看| 三上悠亚av全集在线观看| 9191精品国产免费久久| 9色porny在线观看| 国产人伦9x9x在线观看| av网站免费在线观看视频| 午夜福利一区二区在线看| 精品亚洲乱码少妇综合久久| 在线观看免费视频日本深夜| 大片免费播放器 马上看| 法律面前人人平等表现在哪些方面| 如日韩欧美国产精品一区二区三区| 中文字幕另类日韩欧美亚洲嫩草| 人人妻人人爽人人添夜夜欢视频| 亚洲成人免费电影在线观看| 亚洲中文日韩欧美视频| e午夜精品久久久久久久| 国产精品 国内视频| 在线永久观看黄色视频| 老熟妇仑乱视频hdxx| 99国产精品一区二区三区| 久久久精品94久久精品| 国产熟女午夜一区二区三区| 国产亚洲精品久久久久5区| 久久av网站| 女人高潮潮喷娇喘18禁视频| 水蜜桃什么品种好| 国产精品国产高清国产av | 一夜夜www| 波多野结衣av一区二区av| 亚洲专区中文字幕在线| 激情视频va一区二区三区| 亚洲专区字幕在线| 91精品国产国语对白视频| 国产日韩一区二区三区精品不卡| 多毛熟女@视频| 欧美黄色淫秽网站| 51午夜福利影视在线观看| 热99久久久久精品小说推荐| 在线观看www视频免费| 看免费av毛片| 色婷婷av一区二区三区视频| 三级毛片av免费| 午夜精品久久久久久毛片777| 天天操日日干夜夜撸| 精品人妻在线不人妻| 十八禁高潮呻吟视频| 黄色 视频免费看| 在线 av 中文字幕| a级毛片黄视频| 亚洲午夜精品一区,二区,三区| 91精品国产国语对白视频| 久久国产精品人妻蜜桃| 午夜成年电影在线免费观看| 国产成人欧美在线观看 | 欧美在线一区亚洲| 日韩欧美免费精品| 丁香欧美五月| 老熟妇乱子伦视频在线观看| 精品少妇内射三级| 午夜91福利影院| 9色porny在线观看| 欧美日韩亚洲国产一区二区在线观看 | 亚洲国产中文字幕在线视频| 女人精品久久久久毛片| 免费不卡黄色视频| 国产色视频综合| av在线播放免费不卡| 一边摸一边做爽爽视频免费| 大片电影免费在线观看免费| 国产欧美日韩综合在线一区二区| cao死你这个sao货| 国产一区二区三区在线臀色熟女 | 免费不卡黄色视频| 如日韩欧美国产精品一区二区三区| 在线看a的网站| 亚洲欧美激情在线| 欧美精品av麻豆av| 99香蕉大伊视频| 国产成人一区二区三区免费视频网站| 国产极品粉嫩免费观看在线| 亚洲av日韩精品久久久久久密| videos熟女内射| 美女视频免费永久观看网站| 免费在线观看影片大全网站| 女人被躁到高潮嗷嗷叫费观| 丰满饥渴人妻一区二区三| 12—13女人毛片做爰片一| 十八禁高潮呻吟视频| 国产成人系列免费观看| 91字幕亚洲| 亚洲色图综合在线观看| 99久久精品国产亚洲精品| 国产熟女午夜一区二区三区| 国产欧美日韩综合在线一区二区| 日本一区二区免费在线视频| 最新的欧美精品一区二区| 免费女性裸体啪啪无遮挡网站| 亚洲精品国产精品久久久不卡| 久久精品熟女亚洲av麻豆精品| 极品少妇高潮喷水抽搐| 国产真人三级小视频在线观看| av在线播放免费不卡| 人妻 亚洲 视频| 欧美老熟妇乱子伦牲交| 国产精品.久久久| 久久精品aⅴ一区二区三区四区| 国产欧美日韩一区二区三| 午夜福利,免费看| 色94色欧美一区二区| 在线观看免费视频网站a站| 亚洲一区中文字幕在线| 免费一级毛片在线播放高清视频 | 两性夫妻黄色片| 大香蕉久久成人网| 免费看十八禁软件| 国产精品偷伦视频观看了| 18禁美女被吸乳视频| 亚洲精品中文字幕一二三四区 | 久久久久精品国产欧美久久久| 欧美另类亚洲清纯唯美| 欧美一级毛片孕妇| 一进一出抽搐动态| 中文字幕人妻熟女乱码| 热99re8久久精品国产| 性少妇av在线| 青草久久国产| 日韩大码丰满熟妇| 视频在线观看一区二区三区| 电影成人av| 欧美精品人与动牲交sv欧美| 久久99一区二区三区| 性色av乱码一区二区三区2| 极品教师在线免费播放| 国产精品自产拍在线观看55亚洲 | 久久久欧美国产精品| 国产精品久久久av美女十八| 国产无遮挡羞羞视频在线观看| 亚洲三区欧美一区| 国产一区二区三区视频了| 在线亚洲精品国产二区图片欧美| 丝袜在线中文字幕| 午夜视频精品福利| 老司机靠b影院| av天堂在线播放| 久久久精品免费免费高清| 亚洲成人国产一区在线观看| 国产欧美亚洲国产| 黄频高清免费视频| 亚洲av国产av综合av卡| 麻豆国产av国片精品| 久久精品91无色码中文字幕| 久久中文看片网| av欧美777| 亚洲第一青青草原| 国产精品二区激情视频| 精品国产一区二区三区久久久樱花| 欧美日本中文国产一区发布| 久久中文看片网| 丁香六月天网| 国产精品久久久久久精品古装| 少妇猛男粗大的猛烈进出视频| 色婷婷久久久亚洲欧美| 黄网站色视频无遮挡免费观看| 亚洲国产成人一精品久久久| 国产成人系列免费观看| 亚洲,欧美精品.| 女性生殖器流出的白浆| 亚洲一码二码三码区别大吗| 国产色视频综合| 纯流量卡能插随身wifi吗| 五月天丁香电影| 两性午夜刺激爽爽歪歪视频在线观看 | 成人手机av| 91国产中文字幕| 亚洲人成电影免费在线| 成人国语在线视频| 成人18禁在线播放| 99精品在免费线老司机午夜| 久久精品亚洲精品国产色婷小说| 国产精品1区2区在线观看. | 黑人猛操日本美女一级片| 久久99热这里只频精品6学生| 黑人操中国人逼视频| 国产成+人综合+亚洲专区| 欧美日韩av久久| 一区二区三区激情视频| 国产欧美日韩一区二区三区在线| 精品亚洲乱码少妇综合久久| 亚洲成人免费av在线播放| 午夜福利乱码中文字幕| 欧美变态另类bdsm刘玥| 国产欧美日韩精品亚洲av| 久久人妻熟女aⅴ| 麻豆av在线久日| 亚洲一码二码三码区别大吗| 久久九九热精品免费| 亚洲色图 男人天堂 中文字幕| 老司机福利观看| 久久精品国产综合久久久| 亚洲色图av天堂| 午夜视频精品福利| 久久国产精品男人的天堂亚洲| 久久久水蜜桃国产精品网| 夜夜爽天天搞| 一区二区av电影网| 欧美精品人与动牲交sv欧美| 叶爱在线成人免费视频播放| 国产成人精品无人区| 国产国语露脸激情在线看| 国产一区二区 视频在线| 亚洲午夜精品一区,二区,三区| 手机成人av网站| 亚洲色图综合在线观看| 成人黄色视频免费在线看| 欧美成狂野欧美在线观看| 国产福利在线免费观看视频| 一进一出抽搐动态| 女人精品久久久久毛片| 最近最新中文字幕大全电影3 | 国产成人影院久久av| 日韩熟女老妇一区二区性免费视频| 一本一本久久a久久精品综合妖精| 成人免费观看视频高清| 最近最新中文字幕大全免费视频| 久久精品成人免费网站| 亚洲人成电影观看| 99香蕉大伊视频| 日韩大码丰满熟妇| 中国美女看黄片| 亚洲五月婷婷丁香| 日本wwww免费看| 69精品国产乱码久久久| 最黄视频免费看| 成人手机av| 国产一区二区在线观看av| 亚洲精品乱久久久久久| 婷婷丁香在线五月| 欧美一级毛片孕妇| 亚洲欧洲精品一区二区精品久久久| 中文字幕最新亚洲高清| 亚洲精品av麻豆狂野| a级毛片黄视频| 欧美日韩av久久| 国产精品一区二区精品视频观看| 老司机在亚洲福利影院| 久久久久网色| 岛国在线观看网站| 不卡一级毛片| 婷婷成人精品国产| 少妇精品久久久久久久| 精品少妇黑人巨大在线播放| 精品国产乱码久久久久久男人| 男女免费视频国产| 午夜精品久久久久久毛片777| 男人操女人黄网站| 欧美成人免费av一区二区三区 | 香蕉久久夜色| 咕卡用的链子| 国产一区二区三区视频了| 久久久久久久大尺度免费视频| 不卡一级毛片| 亚洲国产成人一精品久久久| 亚洲美女黄片视频| 亚洲avbb在线观看| 一边摸一边抽搐一进一小说 | 一区二区三区国产精品乱码| 女人爽到高潮嗷嗷叫在线视频| 午夜两性在线视频| 久久婷婷成人综合色麻豆| 波多野结衣av一区二区av| h视频一区二区三区| 欧美老熟妇乱子伦牲交| 国产精品一区二区在线观看99| 亚洲欧美一区二区三区黑人| 久久精品aⅴ一区二区三区四区| tube8黄色片| 考比视频在线观看| 国产成人精品久久二区二区91| 成年女人毛片免费观看观看9 | 精品一区二区三区av网在线观看 | 女同久久另类99精品国产91| 变态另类成人亚洲欧美熟女 | 精品久久久久久久毛片微露脸| 欧美日韩亚洲综合一区二区三区_| 精品久久久久久久毛片微露脸| 亚洲av日韩精品久久久久久密| 久久毛片免费看一区二区三区| 亚洲成人免费av在线播放| 欧美精品亚洲一区二区| 国产精品亚洲一级av第二区| 精品人妻1区二区| 美女高潮喷水抽搐中文字幕| 国产在线视频一区二区| 久久久国产精品麻豆| 亚洲成人国产一区在线观看| 99精品欧美一区二区三区四区| 天堂动漫精品| 丁香欧美五月| 亚洲视频免费观看视频| 久久精品人人爽人人爽视色| 日本a在线网址| 两性夫妻黄色片| 人人妻人人添人人爽欧美一区卜| 91大片在线观看| 成人精品一区二区免费| 一本久久精品| 97人妻天天添夜夜摸| 久久久久久免费高清国产稀缺| 欧美日韩亚洲高清精品| 国产又色又爽无遮挡免费看| 午夜福利在线免费观看网站| 久久人妻熟女aⅴ| 久久中文字幕人妻熟女| 老鸭窝网址在线观看| 黄色片一级片一级黄色片| 国产单亲对白刺激| 亚洲天堂av无毛| 亚洲国产av影院在线观看| 午夜精品久久久久久毛片777| 女人精品久久久久毛片| 精品人妻在线不人妻| 午夜视频精品福利| 人妻一区二区av| 亚洲久久久国产精品| 这个男人来自地球电影免费观看| 亚洲伊人久久精品综合| 国产高清videossex| 少妇猛男粗大的猛烈进出视频| 欧美+亚洲+日韩+国产| 精品视频人人做人人爽| 国产又色又爽无遮挡免费看| 宅男免费午夜| 女人久久www免费人成看片| av超薄肉色丝袜交足视频| 欧美日韩中文字幕国产精品一区二区三区 | 久久久久国内视频| 欧美日韩视频精品一区| 成人国产av品久久久| 国产真人三级小视频在线观看| 国产精品一区二区精品视频观看| 啦啦啦视频在线资源免费观看| 在线看a的网站| 日韩熟女老妇一区二区性免费视频| 国产欧美日韩一区二区三| 亚洲视频免费观看视频| 视频在线观看一区二区三区| 免费在线观看黄色视频的| 2018国产大陆天天弄谢| 国产欧美日韩一区二区精品| 桃花免费在线播放| 欧美另类亚洲清纯唯美| 无限看片的www在线观看| 日韩中文字幕欧美一区二区| 亚洲精品一卡2卡三卡4卡5卡| 日韩欧美国产一区二区入口| 99精品欧美一区二区三区四区| 18禁美女被吸乳视频| 少妇精品久久久久久久| 中文字幕制服av| 亚洲专区字幕在线| 亚洲欧洲精品一区二区精品久久久| 亚洲成人手机| 久久国产精品影院| 成人特级黄色片久久久久久久 | 国产精品自产拍在线观看55亚洲 | 人人妻人人澡人人爽人人夜夜| 下体分泌物呈黄色| 天天操日日干夜夜撸| 久久久久久久久久久久大奶| a级片在线免费高清观看视频| 丝袜美腿诱惑在线| 久久九九热精品免费| 最黄视频免费看| 日本a在线网址| av超薄肉色丝袜交足视频| 久久免费观看电影| 国产精品亚洲av一区麻豆| 波多野结衣一区麻豆| 色尼玛亚洲综合影院| 男男h啪啪无遮挡| 欧美精品亚洲一区二区| av免费在线观看网站| 日韩一区二区三区影片| 巨乳人妻的诱惑在线观看| 免费高清在线观看日韩| 嫁个100分男人电影在线观看| 国产成人免费观看mmmm| 婷婷丁香在线五月| 欧美久久黑人一区二区| 久久精品亚洲熟妇少妇任你| 91九色精品人成在线观看| 激情视频va一区二区三区| 美女午夜性视频免费| 亚洲精品在线观看二区| 天堂动漫精品| 啦啦啦免费观看视频1| 建设人人有责人人尽责人人享有的| 老司机在亚洲福利影院| 免费观看a级毛片全部| 亚洲精品自拍成人| 精品国产一区二区久久| 亚洲精品国产色婷婷电影| 国产精品久久久久成人av| 在线观看66精品国产| 国产一区二区在线观看av| 精品亚洲乱码少妇综合久久| 国产在线一区二区三区精| 十八禁网站免费在线| 国产91精品成人一区二区三区 | 国产91精品成人一区二区三区 | 免费观看av网站的网址| 欧美日韩福利视频一区二区| 久久热在线av| 一级a爱视频在线免费观看| 久久久久久久大尺度免费视频| 亚洲精品久久成人aⅴ小说| 在线永久观看黄色视频| 久久精品熟女亚洲av麻豆精品| 男女床上黄色一级片免费看| 亚洲精品美女久久av网站| videosex国产| 国产精品九九99| 国产av又大| 久久久国产成人免费| 亚洲欧洲精品一区二区精品久久久| 久久国产精品影院| 亚洲成人免费av在线播放| 国产午夜精品久久久久久| a级片在线免费高清观看视频| 性高湖久久久久久久久免费观看| 两性夫妻黄色片| 国产一区二区 视频在线| 黄色怎么调成土黄色| 丰满迷人的少妇在线观看| 九色亚洲精品在线播放| 嫩草影视91久久| 宅男免费午夜| 色精品久久人妻99蜜桃| aaaaa片日本免费| 极品人妻少妇av视频| 亚洲色图综合在线观看| 飞空精品影院首页| 99国产精品一区二区三区| 天天躁日日躁夜夜躁夜夜| 欧美日韩亚洲高清精品| 久久久久久人人人人人| 日韩熟女老妇一区二区性免费视频| 久久性视频一级片| 亚洲欧美日韩高清在线视频 | 97人妻天天添夜夜摸| 久久狼人影院| 亚洲熟妇熟女久久| 99在线人妻在线中文字幕 | 操美女的视频在线观看| 亚洲av片天天在线观看| 亚洲精品av麻豆狂野| 国产日韩欧美亚洲二区| 俄罗斯特黄特色一大片| 最黄视频免费看| 日本av手机在线免费观看| 精品少妇内射三级| 下体分泌物呈黄色| 亚洲国产欧美日韩在线播放| 欧美日本中文国产一区发布| 亚洲av电影在线进入| 69精品国产乱码久久久| 50天的宝宝边吃奶边哭怎么回事| 精品国产超薄肉色丝袜足j| 妹子高潮喷水视频| 日韩精品免费视频一区二区三区| 久久久水蜜桃国产精品网| 国产极品粉嫩免费观看在线| 国产在线视频一区二区| 亚洲成人免费av在线播放| 九色亚洲精品在线播放| 日韩欧美一区视频在线观看| 久久人妻福利社区极品人妻图片| 18禁黄网站禁片午夜丰满| 亚洲国产精品一区二区三区在线| 99久久人妻综合| 国产成人精品在线电影| a级毛片在线看网站| 国产精品偷伦视频观看了| 久久久欧美国产精品| 麻豆成人av在线观看| 国产免费av片在线观看野外av| 色精品久久人妻99蜜桃| 性少妇av在线| 交换朋友夫妻互换小说| 欧美在线一区亚洲| 热99re8久久精品国产| 搡老乐熟女国产| 9191精品国产免费久久| 黄网站色视频无遮挡免费观看| 国产成人欧美| 曰老女人黄片| 日韩欧美国产一区二区入口| 日本精品一区二区三区蜜桃| 国产免费福利视频在线观看| 制服诱惑二区| 欧美成人免费av一区二区三区 | 亚洲精品在线观看二区| 国产日韩欧美亚洲二区| 一级,二级,三级黄色视频| 久久精品国产综合久久久| 亚洲精品美女久久av网站| svipshipincom国产片| 国产亚洲一区二区精品| 啪啪无遮挡十八禁网站| 黄色怎么调成土黄色| 美女视频免费永久观看网站| 欧美精品人与动牲交sv欧美| www.熟女人妻精品国产| 国产亚洲一区二区精品| 精品国产一区二区三区久久久樱花| 精品福利永久在线观看| 精品国内亚洲2022精品成人 | 精品高清国产在线一区| 在线 av 中文字幕| 日韩欧美一区视频在线观看| a级毛片在线看网站| 国产av国产精品国产| 日韩一区二区三区影片| 满18在线观看网站| 国产av又大| 成在线人永久免费视频| 亚洲伊人色综图| 免费在线观看完整版高清| 欧美精品一区二区免费开放| 亚洲国产中文字幕在线视频| 午夜福利在线观看吧| 99精品久久久久人妻精品| 99热网站在线观看| 成人精品一区二区免费| 中文字幕最新亚洲高清| 国产免费视频播放在线视频| 中文字幕制服av| 午夜福利,免费看| 亚洲性夜色夜夜综合| 最新在线观看一区二区三区| 色播在线永久视频| 国产精品影院久久| 免费高清在线观看日韩| 精品久久久久久电影网| 欧美精品啪啪一区二区三区| 999久久久国产精品视频| 成人亚洲精品一区在线观看| 日本vs欧美在线观看视频| 色尼玛亚洲综合影院| 亚洲成人免费av在线播放| 18禁美女被吸乳视频| 久久性视频一级片| 两性夫妻黄色片| 一本综合久久免费| 久久久久久久大尺度免费视频| 18禁美女被吸乳视频| av网站免费在线观看视频| 久久影院123| 国产男靠女视频免费网站| 亚洲七黄色美女视频| 久久国产亚洲av麻豆专区| 亚洲精品成人av观看孕妇| 狠狠精品人妻久久久久久综合| 黑丝袜美女国产一区| www.熟女人妻精品国产| 欧美 亚洲 国产 日韩一| 国产在线一区二区三区精| 1024视频免费在线观看| av国产精品久久久久影院| 久久影院123| 欧美日韩亚洲高清精品| 国产成人精品无人区| 国产av国产精品国产| 成年动漫av网址| 亚洲国产精品一区二区三区在线| 日日爽夜夜爽网站|