• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Deep Learning Approach for Prediction of Protein Secondary Structure

    2022-08-24 07:02:24MuhammadZubairMuhammadKashifHanifEatedalAlabdulkreemYazeedGhadiMuhammadIrfanKhanMuhammadUmerSarwarandAyeshaHanif
    Computers Materials&Continua 2022年8期

    Muhammad Zubair,Muhammad Kashif Hanif,*,Eatedal Alabdulkreem,Yazeed Ghadi,Muhammad Irfan Khan,Muhammad Umer Sarwar and Ayesha Hanif

    1Department of Computer Science,Government College University,Faisalabad,Pakistan

    2Department of Computer Science,College of Computer and Information Sciences,Princess Nourah Bint Abdulrahman University,Riyadh 11671,Saudi Arabia

    3Department of Software Engineering/Computer Science,Al Ain University,Abu Dhabi,United Arab Emirates

    Abstract: The secondary structure of a protein is critical for establishing a link between the protein primary and tertiary structures.For this reason,it is important to design methods for accurate protein secondary structure prediction.Most of the existing computational techniques for protein structural and functional prediction are based on machine learning with shallow frameworks.Different deep learning architectures have already been applied to tackle protein secondary structure prediction problem.In this study,deep learning based models,i.e.,convolutional neural network and long short-term memory for protein secondary structure prediction were proposed.The input to proposed models is amino acid sequences which were derived from CulledPDB dataset.Hyperparameter tuning with cross validation was employed to attain best parameters for the proposed models.The proposed models enables effective processing of amino acids and attain approximately 87.05% and 87.47% Q3 accuracy of protein secondary structure prediction for convolutional neural network and long short-term memory models,respectively.

    Keywords: Convolutional neural network;machine learning;protein secondary structure;deep learning;long short-term memory;protein secondary structure prediction

    1 Introduction

    Proteins are very complex molecules found in all living organisms.Proteins act as metabolites in chemical reactions,chemical messengers,or hormones in inner contact and transportation pathways such as oxygen delivery in the blood.Proteins are also engaged in the storage absorption of material,creating complicated structures,deoxyribonucleic acid (DNA) replication,reacting to stimuli,providing shape to cells and animals,catalyzing metabolic events,and conveying chemicals and the conservation of systems.Proteins are composed of amino acids,chemical molecules with amine(NH2)and carboxyl(COOH) functional groups.Proteins are polymeric linear chains constituted of amino acids.

    To assess protein activity at the molecular and cellular levels,it is necessary to find the arrangement of a specific sequence.Therefore,it has become more critical to predict the proteins structures from their primary sequence in bioinformatics.The three-dimensional structure determines its nature and role in its environment.The three-dimensional structure of a protein can help find the vast range of functions of individual proteins.This is why understanding the protein structure is the first step toward recognizing the function of a newly identified protein[1].The interactions between protein folding into a complicated three-dimensional form are due to amino acids that,under certain conditions,remain constant[2].This is one of the most difficult problems in bioinformatics.

    The secondary structure may be thought of as an information bridge that connects the primary sequence and the tertiary structure.The primary structure of millions of proteins is well understood.However,the secondary and tertiary structures for the vast majority of proteins remain unclear.Moreover,only a limited fraction of proteins has secondary and tertiary structures.Therefore,protein structure and function study can further enhance nutrition supplements,medications and antibiotics[3].

    Furthermore,the analysis of existing proteins will aid in treating diseases and treating various biological problems.The most crucial problems are money,time,and competence to predict Protein Secondary Structure(PSS)from an experimental perspective.Protein structures can be predicted using crystallography and NMR[4],which takes highly specialized knowledge,a high level of talent,and a lot of fortune.One prediction approach is ab-initio prediction[5]which attempts to forecast protein structure solely based on the primary structure and ignores any trends.Chothla and Levitt published the first Protein Secondary Structure Prediction(PSSP)technique in 1976.

    ML algorithms,Bayesian statistics,nearest neighbor and established sequence-to-structure experimentation are all examples of approaches that can be used to explore and forecast biological patterns.Accelerated evolution of proteomics and genomics technologies for sequencing of protein and DNA have culminated in an immense increase of proteins sequence data.Determining protein structure entails a series of computational activities,with secondary structure prediction being a crucial first step[6].Various sorts of factors,including geometrical,physicochemical,and topological factors,can be used to determine PSS.Finding protein secondary and tertiary structures from their chain sequence is challenging.

    Protein structure prediction technologies have been divided into three generations [7].The first generation emerged before the 1980s.The accuracy of these methods was below 60%.Chou-Fasman’s method is one of these methods.The second generation appeared between 1980 and 1992.These approaches might increase prediction accuracy to some amount.However,the total accuracy was under 65%.After 1992,the third generation of techniques developed,which usually employed multiple sequence alignment files to input an advanced Machine Learning(ML)model to predict PSS.PHD and PSIPRED were the typical techniques and the total accuracy of this generation was approximately between 76%to 80%[8].

    Many ML approaches have been developed to forecast secondary structure and demonstrated good progress through evolutionary awareness and statistical information on amino acid subsequences.This study employed deep learning-based Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) models to predict PSS.Stratified k fold cross-validation and hyperparameter tuning were used to obtain the best parameters for these models.Then,these models were retrained using optimized parameters to attain better performance.

    The rest of the paper is divided into various sections.Section 2 discusses the background of proteins and Deep Learning (DL).Related works are presented in Section 3.The datasets and the methodology are described in Section 4.Section 5 evaluates the performance of the proposed techniques.Finally,Section 6 concludes this study.

    2 Background

    Protein functionality is determined by the amino acids that make up the protein.This depends on how these molecules fold across space,assemble and work.Protein functionality can help researchers better understand why people age,why they get sick from harmful viral diseases(like cancer),how to find a cure for a disease(like the cure for covid-19)and other‘tough’questions.The roles of proteins are linked to their composition,which is influenced by physicochemical parameters.Determining a protein’s native structure in solution is the same as figuring out how the protein can fold.The protein folding issue has produced a great deal of knowledge about the processes that govern how this process occurs,which physical and chemical interactions have the most significant impact,and how the amino acid sequence of a protein stores details about its structure [9].In general,proteins fold rapidly to their native state,while environmental factors such as extreme temperatures or pH can prevent or reverse this process (protein denaturing).Furthermore,specific proteins use chaperones to prevent premature misfolding or unwanted aggregation during synthesis [10].Secondary structures fold to form temporary local conformations maintained by the evolving tertiary structure[11].

    The PSSP problem needs to be addressed since the PSS can help predict the tertiary structure that contains details about a protein’s functions.There exist millions of proteins.However,only a limited percentage of recognized proteins have been studied because the experimental techniques for determining the tertiary structure of proteins are expensive.For this reason,PSSP can be used to identify a protein’s tertiary structure with greater precision and less effort.

    2.1 Amino Acids

    Proteins are composed of amino acids,which are the crucial building components.An amino acid is a molecule that consists of a carbon atom with the primary amino group(NH2),a carboxylic acid group(COOH),a side chain(R group),and an H atom linked to it.The essential components of an amino acid are carbon (C),hydrogen (H),oxygen (O),and nitrogen (N).On the other hand,other components can be found in the side chains of specific amino acids(Fig.1).Each amino acid can be expressed by one of three characters from the English alphabet.A sequence of characters can be used to describe a specific amino acid.No matter how minor,any modification in this series will result in a different protein with distinct properties and functions[12].

    Figure 1:Amino acid structure[12]

    Amino acids can be categorized based on the location of the significant structural groups(i.e.,alpha (α),beta (β),gamma (γ),and delta (δ)),pH level and side-chain group type [13].In addition to neurotransmitter transport and biosynthesis,amino acids are involved in several other processes.Peptides are short chains of amino acids (30 or fewer) joined by peptide bonds,whereas polypeptides are long,continuous,and unbranched peptide chains.Proteins are made up of one or more polypeptides[14].

    2.2 Proteins Structures

    A hierarchical methodology can be used to analyze the composition of proteins and study their various shapes more accurately [15].The protein structures are classified into four categories,i.e.,primary,secondary,tertiary,and quaternary.The primary structure is commonly a generalized form of amino acids arranged within the unfolded protein.A protein’s sequence determines its structure and function,which is unique to that protein.For example,Insulin is made up of 51 amino acids in two peptide chains,i.e.,A chain has 21 amino acids and B chain has 30 amino acids[16].

    The local conformation of a polypeptide backbone is known as the secondary structure of proteins.The secondary structure is characterized by patterns of hydrogen bonds between the main chain peptide groups and refers to the local folding of the polypeptide in certain areas.The secondary structure is divided into theα-helix and theβ-strand orβ-sheets categories[17].The three-dimensional structure of monomeric and multimeric protein molecules is called the tertiary structure of proteins.Ionic bonding,hydrophobic interactions,hydrogen bonding,and disulfide linkages are among the chemical interactions that define the structure of a polypeptide chain.Interactions between R groups give the protein its complex three-dimensional tertiary structures.The quaternary structure of a protein is a three-dimensional structure made up of two or more individual polypeptide chains that work together as a single functional unit.Insulin,for example,is a globular protein with a mixture of hydrogen and disulfide bonds that allow it to clump into a ball shape.

    2.3 Deep Learning

    DL is a subfield of ML that employs a layered framework to learn representations.Each layer of the layered architecture operates with more meaningful and sophisticated representations.The term‘deep’in DL refers to a number of successive levels of expression present in a model.Each layer can operate on the output generated by the previous layer.Thus,each layer operates by detecting the pattern over the last layer’s output.This makes the corresponding layer work with more complex,advanced,and significant pattern recognition than the previous layers.This makes DL models perfect for dealing with complicated data.There are several various types of DL models.Each of these models is optimal for different types of data and computational problems.For example,CNN’s are good for image processing,whereas Recurrent Neural Networks (RNNs) are good for natural processing language.DL approaches have been widely used in various areas,including bioinformatics[18-22].DL techniques have been applied to perform sequence analysis[20],structure prediction and reconstruction[21],as well as biomolecular property and functionality prediction[22].

    3 Related Work

    The development of DL techniques revealed the elegant applications of ML for many application domains.DL techniques are being used to study protein sequence,structures,and functions.This section presents the work related to existing applications of DL techniques in protein analysis and existing computational approaches for predicting protein structure and functional characteristics.Yang et al.(2013) introduced a new approach employing a large margin nearest neighbor method(Tab.1) for the PSSP problem.Experimental results revealed that the proposed approach achieved greater prediction accuracy than previous nearest neighbor models.For RS126 and CB513 datasets,Q3accuracy is 75.09%and 75.44%,respectively[23].Feng et al.(2014)used increasing diversity blended with a quadratic discriminant approach to predict the composition of core residues.For 20 amino acid residues,the precision of predicted secondary structures varies from 81%to 88%[24].

    Table 1:Summary of related work for PSSP

    Spencer et al.(2015)have created a PSS predictor called DNSS that uses the position-specific score matrix produced by DL network architectures and PSI-BLAST.This methodology was based on a DL network and used to forecast secondary structure for a completely autonomous research dataset for 198 proteins with a Q3accuracy of 80.7%[25].Heffernan et al.(2015)used DL architecture in three loops.As a result,they attained 82%accuracy for a dataset consisting of 1199 proteins[26].

    Nguyen et al.(2015) introduced a system called MOIT2FLS for PSSP using the quantization method of adaptive vectors for each type of secondary structure to create an equivalent number of simple rules.The genetic algorithm was used to change the MOIT2FLS parameters optimally [27].Experimental findings indicate that the proposed solution dominates the conventional approaches of artificial neural network models,Chou-Fasman method and Garnier-Osguthorpe-Robson method.Zamani et al.(2015) presented a PSS classification algorithm using genetic programming using IF regulations for a multi-target classification task.The experiments were done on two datasets,RS126 and CB513,to attain Q3accuracy of 65.1%and 66.4%,respectively[29].

    Heffernan et al.(2017) proposed Bidirectional Recurrent Neural Networks (BRNNs).The proposed model was capable of recording long-range experiences without using a window.The Q3accuracy of 83.9% was achieved for TS115 [28].Asgari et al.(2019) developed a software“DeepPrime2Sec”using DL method CNN-BiLSTM network to predict PSS from primary structure.The expected structure was identical to the target structure even though the PSS’s exact nature could not be predicted.For eight classes of PSS,approximately 70.4%accuracy was obtained for the CB513 dataset using ensemble top-k models[31].Li et al.(2019)built an ensemble model based on Bi-LSTM for PSSP.The proposed model was tested using three separate datasets,i.e.,CB513,data1199,and CASP203 proteins.Ensemble model achieved 84.3%Q3accuracy and 81.9%SOV score using 10-fold cross-validation[33].

    Cheng et al.(2020) proposed a method based on LSTM and CNN.Cross-validation tests were conducted on a dataset of 25pdb and achieved 80.18%accuracy,which was better than using a single model[30].Ratul et al.(2020)implemented a deep neural network model called PS8-Net to increase the accuracy of eight-class PSSP.For the CullPdb6133,CASP11,CASP10 and CB513 datasets,the proposed PS8-Net achieved 76.89%,71.94%,76.86%and 75.26%Q8accuracy,respectively[32].

    4 Methodology

    PSSP depends on protein data,access to protein databanks,and secondary structure information for known sequences.Proteins and their structures are being found slowly but steadily through exclusion chromatography,mass spectroscopy and nuclear resonance spectroscopy[34].Fig.2 shows the proposed methodology.This section describes the datasets used,data preprocessing techniques,and the proposed deep learning models.

    Figure 2:Proposed methodology

    4.1 Dataset

    Proteins are discovered and inserted into protein databanks such as the RCSB Protein Data Bank(PDB).This data contains protein names,lengths,structures(primary,secondary,tertiary and quaternary),and other biological facts.In this study,the CulledPDB dataset from the PISCES server is used.PISCES[35]is a public server that selects protein sequences from the PDB based on sequence identity and structural quality criteria.PISCES can give lists chosen from the complete PDB or usersupplied lists of PDB entries or chains.PISCES produces much better lists than BLAST servers,which cannot recognize many associations with less than 40% sequence identity and frequently overstate sequence identity by matching only well-conserved fragments[35].CulledPDB datasets on the PISCES service offer the most comprehensive list of high-resolution structures that meet the sequence identity and structural quality cut-offs.After downloading PISCES,we removed peptides with high similarity.The dataset has

    ? pdb_id:the id that was used to find its entry

    ? chain code:The chain code is required to find a specific peptide(chain)in a protein that contains numerous peptides(chains).

    ? seq:the peptide’s sequence of amino acids

    ? sst8:Eight state(Q8)secondary structure

    ? sst3:Three state(Q3)secondary structure

    ? len:the number of amino acids in the peptide

    ? hasnonstdaa:whether there are any non-standard amino acids in the peptide(i.e.,B,O,U,X,or Z).

    4.2 Data Preprocessing

    The dataset is in text/categorical form.Many ML methods cannot perform well if data is in text/categorical form.The proposed models need data in numerical form.Therefore,data preprocessing techniques were applied to deal with missing values,noise in data,standardization of data,textual data,and categorical variables.There are several methods to convert categorical variables to numerical numbers.Many ML algorithms cannot work with categorical variables effectively.The categories must be numerically transformed.Neural network are used in DL to predict numerical input values.Each method has its own set of trade-offs and effects on the feature set.In this study,the one-hot encoding technique is used.This encoder is part of the SciKit-learn toolkit (a popular Python library) and transforms text or categorical data into numerical data.Single hot encoding for the outcome variable might result in a more complex set of estimations when a single label is used.This work deals with the data in text form and needs to be changed in numerical form.Protein sequences are converted from text to numerical data using one-hot encoding.For example,ELVDNAVGGDL,LVDNAVG-GDLS,VDNAVGGDLSK,VETRTIVRFNR,and RIFIIIVGSLI are samples of some structures from the dataset used in this research (Fig.3).These structures are taken from the PDB website and have all the information.These are the samples of some protein structures used in the dataset.The raw dataset has 393733 structures of proteins.Each protein has its unique id.For example,Figs.3a and 3b have id 1GMC and 2F9N,respectively.

    Figure 3:Sample protein crystal structure(a)1GMC1https://www.rcsb.org/structure/1GMC(b)2F9N2https://www.rcsb.org/structure/2F9N

    Tokenization is defined as separating a large amount of text into smaller chunks known as tokens.These pieces or tokens are precious for finding patterns.Tokenization also allows for the replacement of sensitive data components with non-sensitive ones.Tokenization can be performed for either individual words or entire phrases.Stop words in the text add no sense to the phrase.Removing stop words will not influence text processing for the specified goal.They are deleted from the lexicon to minimize noise and the size of the feature set.In this study,input_grams is a list containing windows of variable length for each sequence.Each window is a ‘word’and is encoded as an integer using a dictionary.Each list of window encoded integers is padded with 0 s until it becomes 128 integers long.

    For example,the sequence KCK will have three frames:KCK,CK and K.In this case,preferred_amino_acid_chunk_size is 3.Those frames are then converted into integers and added to a list for the sequence added to input_data.The target_data contains the one-hot encoded secondary structure with an additional integer for no structure used when there is no sequence in the padding.The target_data for sequence KCK could thus possibly be [[0.1.0.0.],[0.0.0.1.],[0.1.0.0.],[1.0.0.0.]...].After tokenization,the data will have 77629 different windows and four different possible structures.

    4.3 Deep Learning Models

    This study employed CNN and LSTM deep learning models for the PSSP problem.These models were trained on the preprocessed dataset to attain better performance.To evaluate the performance for each parameter combination,cross-validation was used.The training and validation dataset is used to find accuracy for each parameter setting for each split in the cross-validation.Stratified k fold cross-validation with hyperparameter tuning is used to find the best parameters[36,37].The proposed models were retrained using these best parameters.We considered batch size,a number of the epoch,and activation function at each layer.The best parameters values attained using hyperparameter tuning for epochs,batch_size,and optimizer attained are 20,128,and Adam,respectively.Furthermore,dimensionality reduction techniques were employed[38].

    The proposed CNN model for PSSP is shown in Fig.4.The input to the model is protein images.First,dictionary integer input is converted to floats by the embedding layer.This layer creates a filter(also known as a feature detector).The neural network can learn features depending on filters.We have employed 128 filters and 40 x 128 neuron matrix is the output of the first layer.The weights of one single filter are stored in each column of the output matrix.Each filter will have 40 weights if the kernel size is defined and the input matrix length is considered.Next,a dropout layer is used to prevent over-fitting(deactivates 30%of input neurons).The dropout layer assigns zero weights to the network’s neurons at random.In this study,we set a rate of 0.3 which means less than half of the neurons will have no weight.The network becomes less sensitive to tiny fluctuations in the data due to this process.It should improve the accuracy of previously unknown data.The output of this layer is a 40 x 64 neuron matrix.Then,another convolutional layer is added that now has 64 filters.Next,another dropout layer is added that deactivates 30% of input neurons.The final convolution layer reduces output to four classes and has a softmax activation function(Fig.4).

    Figure 4:Proposed CNN model layers for PSSP

    Next,the LSTM model was employed for PSSP (Fig.5).The embedding layer converts data sequences into float values.A bidirectional LSTM (BiLSTM) layer is used to learn long-term bidirectional relationships between time steps of time series or sequence data.These dependencies may be helpful when the network learns from the whole time series at each time step.This model added a bidirectional layer that will pass on information from the past and future states to the output.

    Figure 5:Proposed LSTM model for PSSP

    Moreover,the time distribution layer is added that will apply the dense layer to every neuron and thus make the output“n_tags.”The time distribution layer does not affect how layers work.Instead,the goal is to create a second “time”dimension (which may or may not represent time).With this temporal dimension in mind,the wrapped layer is applied to each slice of the input tensor.

    5 Results and Discussion

    The attempt to solve the PSSP problem was significant since the experimental methods that are currently available are extremely expensive in both money and time.PSSP focuses on predicting structures as correctly as feasible based on their basic structure.The Q3metric is frequently employed for PSSP to determine the effectiveness of the trained models.The Q3accuracy is the percent of correctly predicted protein secondary structure using three states and is calculated by dividing the number of adequately categorized amino acids by the total amount of amino acids.In this study,per residue Q3accuracy metric is used to evaluate the performance of the proposed models.

    This research aimed to design and implement CNN and LSTM models that can predict the PSSP from its primary structure.The dataset used in this research was a combination of PDB and PISCES.The dataset was split into training and testing test data.40%of sequences are selected for testing and 60% for training the model.Fig.6 shows sample training and testing data.The proposed model is trained on protein to check how accurately the model reads proteins structures like helix,sheet,and loop.The portion which the model can not predict is labeled as unknown.

    Figure 6:Sample dataset used for training and testing

    The proposed models were trained and tested on the prepared dataset.This dataset lists chains of proteins in rows.The chains are identified by a chain code and the underlying protein id within the protein database.In addition to pdb id and chain codes,the dataset also has the sequence of amino acids and the secondary structures (3 and 8 states) for a given chain.The categorical cross-entropy is reduced by training the model.The Q3accuracy is obtained by computing the accuracy exclusively for coding characters.Cross validation was employed to address overfitting and underfitting issues.The loss function is one of the most important aspects of neural networks.The term“l(fā)oss”refers to a prediction error made by a neural network.The loss function determines the gradients.Gradients are also used to update the neural network weights.Maximum Q3accuracy obtained using CNN and LSTM is 87.05%and 87.47%,respectively.

    Fig.7 shows the CNN training and testing accuracy.Maximum accuracy for CNN’s training and validation is approximately 97.50% and 93.23%,respectively.The training and validation accuracy of LSTM is 97.61%and 93.44%,respectively(Fig.8).Testing accuracy is less than training accuracy for CNN and LSTM.The reason is model was trained using training data.However,testing data is a collection of data which is new to the model.The accuracy of both models is approximately 90%which shows model performs well on dataset.This is due to choosing best parameters obtained using hyperparameter tuning and cross validation.The results demonstrate that features derived from CNN and LSTM models can significantly enhance the accuracy of PSSP.

    Figure 7:Performance of the proposed CNN model

    Figure 8:Performance of the proposed LSTM model

    6 Conclusion

    PSS provides important characteristics for predicting protein tertiary structure.However,PSSP prediction techniques used in laboratories are expensive and time consuming.In this work,CNN and LSTM models were proposed to predict PSS from the amino acid sequences.The input to proposed models was amino acids which were obtained from the CulledPDB dataset.Moreover,this study employed cross-validation with hyperparameter tuning to enhance the performance of the proposed models.Experimental results showed proposed CNN and LSTM models achieved 87.05%and 87.47%Q3accuracy,respectively.Despite the goodness and validity of proposed methods,current methods cannot deal with highly long dependencies.In future work,we will apply the attention mechanism to the study of low-frequency long-range interactions of PSSP.

    Funding Statement:Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2022R161),Princes Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲精品中文字幕在线视频 | 黄色视频在线播放观看不卡| 日韩成人av中文字幕在线观看| 国产国拍精品亚洲av在线观看| 日本色播在线视频| 成人毛片60女人毛片免费| 日本熟妇午夜| 人妻 亚洲 视频| 亚洲久久久久久中文字幕| 最近的中文字幕免费完整| videossex国产| 国产综合精华液| 成人漫画全彩无遮挡| 久久久国产一区二区| 一级毛片 在线播放| 日韩中字成人| 久久久午夜欧美精品| 欧美三级亚洲精品| 最近中文字幕2019免费版| 特级一级黄色大片| 中文资源天堂在线| 国产探花在线观看一区二区| 丝瓜视频免费看黄片| 最近2019中文字幕mv第一页| 美女内射精品一级片tv| 黄色怎么调成土黄色| 国产女主播在线喷水免费视频网站| 白带黄色成豆腐渣| 亚洲精华国产精华液的使用体验| 国产色婷婷99| 新久久久久国产一级毛片| 免费黄网站久久成人精品| 国产乱来视频区| 国产成人aa在线观看| 99久久精品国产国产毛片| 欧美xxxx黑人xx丫x性爽| 国产成人免费无遮挡视频| 欧美性感艳星| 日本猛色少妇xxxxx猛交久久| 亚洲美女视频黄频| eeuss影院久久| 欧美日韩在线观看h| 日产精品乱码卡一卡2卡三| av网站免费在线观看视频| 午夜福利网站1000一区二区三区| 国产高清不卡午夜福利| 久久久久久久久久久免费av| 欧美+日韩+精品| 亚洲人成网站在线观看播放| 免费人成在线观看视频色| 国产一区有黄有色的免费视频| 中文在线观看免费www的网站| 亚洲不卡免费看| 中文字幕免费在线视频6| 欧美成人精品欧美一级黄| 久热久热在线精品观看| 亚洲精品中文字幕在线视频 | 日韩一区二区视频免费看| 免费少妇av软件| 美女被艹到高潮喷水动态| 成人美女网站在线观看视频| 男女那种视频在线观看| 国产亚洲av片在线观看秒播厂| 国产精品久久久久久久久免| 亚洲精品国产av蜜桃| 精品亚洲乱码少妇综合久久| 日本三级黄在线观看| 国产日韩欧美亚洲二区| 成年av动漫网址| 特大巨黑吊av在线直播| 欧美极品一区二区三区四区| 视频中文字幕在线观看| 国产亚洲最大av| 欧美日韩在线观看h| 中文欧美无线码| 亚洲天堂av无毛| 久久精品国产自在天天线| 久久久久精品久久久久真实原创| 26uuu在线亚洲综合色| 黄片无遮挡物在线观看| 亚洲天堂av无毛| 国产精品秋霞免费鲁丝片| 我的女老师完整版在线观看| 亚洲自偷自拍三级| 亚洲内射少妇av| 久久精品综合一区二区三区| 成人特级av手机在线观看| 久久综合国产亚洲精品| 国产av国产精品国产| 在线观看免费高清a一片| 一级毛片我不卡| 能在线免费看毛片的网站| 在现免费观看毛片| 国产片特级美女逼逼视频| 老司机影院成人| 欧美成人a在线观看| 国精品久久久久久国模美| 少妇猛男粗大的猛烈进出视频 | 精品久久久久久久人妻蜜臀av| 熟女av电影| 亚洲在久久综合| 国产精品.久久久| av天堂中文字幕网| 寂寞人妻少妇视频99o| 国产高潮美女av| 亚洲精品日韩在线中文字幕| 一区二区三区精品91| 精品国产三级普通话版| 中国三级夫妇交换| 我的女老师完整版在线观看| 大码成人一级视频| 我要看日韩黄色一级片| 最近中文字幕高清免费大全6| 一区二区三区免费毛片| 亚洲av免费高清在线观看| 亚洲欧美中文字幕日韩二区| 久久精品久久久久久久性| 国产精品伦人一区二区| 丝瓜视频免费看黄片| 日韩视频在线欧美| 欧美+日韩+精品| 老司机影院毛片| 亚洲欧美精品自产自拍| 久久久久久九九精品二区国产| av在线观看视频网站免费| 久久99精品国语久久久| 亚洲成人精品中文字幕电影| 精品国产三级普通话版| 大片电影免费在线观看免费| 国产亚洲最大av| 一区二区三区四区激情视频| 一本久久精品| 在线 av 中文字幕| 美女视频免费永久观看网站| 老女人水多毛片| 国产欧美日韩精品一区二区| 特级一级黄色大片| 最近中文字幕2019免费版| 日韩av不卡免费在线播放| 亚洲欧美精品专区久久| 亚洲美女视频黄频| 国国产精品蜜臀av免费| av在线天堂中文字幕| 国产免费视频播放在线视频| 国内精品宾馆在线| 一级黄片播放器| 高清毛片免费看| 男女那种视频在线观看| 欧美老熟妇乱子伦牲交| 欧美激情国产日韩精品一区| 欧美区成人在线视频| 欧美精品人与动牲交sv欧美| 97人妻精品一区二区三区麻豆| 亚洲av男天堂| 男女国产视频网站| 亚洲aⅴ乱码一区二区在线播放| 18禁裸乳无遮挡免费网站照片| 麻豆成人午夜福利视频| 国产精品一二三区在线看| 成人亚洲精品av一区二区| 观看免费一级毛片| 秋霞伦理黄片| 久久久久久久久久久免费av| 国产免费一级a男人的天堂| 国产黄频视频在线观看| 亚洲国产av新网站| 一边亲一边摸免费视频| 久久精品久久精品一区二区三区| 亚洲av免费高清在线观看| 久久精品久久久久久久性| 亚洲国产精品国产精品| 80岁老熟妇乱子伦牲交| 国产精品一区二区在线观看99| 成人毛片a级毛片在线播放| 国产一区二区在线观看日韩| 久久人人爽人人爽人人片va| 校园人妻丝袜中文字幕| 春色校园在线视频观看| 中文字幕制服av| 精品一区二区三卡| 天天一区二区日本电影三级| 高清午夜精品一区二区三区| 亚洲精品久久久久久婷婷小说| 婷婷色综合大香蕉| 男女国产视频网站| 91精品伊人久久大香线蕉| 国产老妇伦熟女老妇高清| 嘟嘟电影网在线观看| 国产色婷婷99| 久久久久国产网址| 国产成人精品久久久久久| 永久网站在线| 亚洲精品自拍成人| 国产精品一及| 我要看日韩黄色一级片| 欧美xxxx黑人xx丫x性爽| 麻豆精品久久久久久蜜桃| 丝袜喷水一区| 国产毛片在线视频| 免费观看av网站的网址| 日韩欧美精品v在线| 日韩av免费高清视频| 啦啦啦中文免费视频观看日本| 男人和女人高潮做爰伦理| 香蕉精品网在线| 日韩视频在线欧美| 一个人看的www免费观看视频| 国产女主播在线喷水免费视频网站| 哪个播放器可以免费观看大片| 国产日韩欧美亚洲二区| 免费看不卡的av| 亚洲欧美精品自产自拍| 美女主播在线视频| 国产亚洲午夜精品一区二区久久 | 91精品一卡2卡3卡4卡| 肉色欧美久久久久久久蜜桃 | av女优亚洲男人天堂| 女的被弄到高潮叫床怎么办| 看黄色毛片网站| 在线天堂最新版资源| 免费观看a级毛片全部| 夫妻午夜视频| 国语对白做爰xxxⅹ性视频网站| 亚洲成人一二三区av| 秋霞伦理黄片| 久久久久久久久久人人人人人人| 中文欧美无线码| 少妇的逼好多水| 能在线免费看毛片的网站| 国产v大片淫在线免费观看| 美女被艹到高潮喷水动态| 九九在线视频观看精品| 一区二区三区精品91| 国产男人的电影天堂91| 精品亚洲乱码少妇综合久久| 国产精品人妻久久久久久| 亚洲欧美一区二区三区黑人 | 免费观看av网站的网址| 欧美日韩视频精品一区| 亚洲自偷自拍三级| 久久韩国三级中文字幕| 99热全是精品| 欧美97在线视频| 欧美激情久久久久久爽电影| av在线观看视频网站免费| 国产欧美另类精品又又久久亚洲欧美| 青青草视频在线视频观看| 水蜜桃什么品种好| 男女那种视频在线观看| 久久久久久久久久久免费av| freevideosex欧美| av免费在线看不卡| 国产成人免费无遮挡视频| 国产一级毛片在线| 日韩一区二区视频免费看| 国产精品一区二区性色av| 色吧在线观看| 69av精品久久久久久| 韩国高清视频一区二区三区| 亚洲人成网站在线观看播放| 亚洲精品亚洲一区二区| 国产欧美另类精品又又久久亚洲欧美| 精品久久久久久久末码| 中文字幕制服av| 伊人久久精品亚洲午夜| 中文欧美无线码| 美女xxoo啪啪120秒动态图| 午夜爱爱视频在线播放| 观看免费一级毛片| 亚洲成色77777| 91aial.com中文字幕在线观看| 舔av片在线| 欧美xxxx性猛交bbbb| 国产欧美日韩精品一区二区| 亚洲欧美一区二区三区黑人 | xxx大片免费视频| 国产精品久久久久久av不卡| 99热这里只有精品一区| 国产成人午夜福利电影在线观看| 日本免费在线观看一区| 亚洲精品日韩av片在线观看| 久久影院123| 免费黄色在线免费观看| 黄色欧美视频在线观看| 三级国产精品欧美在线观看| 男女下面进入的视频免费午夜| 日韩强制内射视频| 亚洲成人久久爱视频| 国产色爽女视频免费观看| 在线a可以看的网站| av国产免费在线观看| 少妇裸体淫交视频免费看高清| 99热全是精品| 人人妻人人看人人澡| 亚洲精品久久久久久婷婷小说| 久久精品综合一区二区三区| 亚洲精品日本国产第一区| 欧美一级a爱片免费观看看| 伊人久久精品亚洲午夜| 美女主播在线视频| 国产精品女同一区二区软件| 韩国av在线不卡| 在线播放无遮挡| freevideosex欧美| 国产精品精品国产色婷婷| 亚洲精品第二区| 人人妻人人看人人澡| 欧美精品人与动牲交sv欧美| 国产精品国产三级国产av玫瑰| 一级二级三级毛片免费看| 啦啦啦啦在线视频资源| 另类亚洲欧美激情| 在线观看av片永久免费下载| 国产精品三级大全| 久久午夜福利片| 蜜桃久久精品国产亚洲av| 不卡视频在线观看欧美| 国产真实伦视频高清在线观看| 亚洲最大成人av| 狂野欧美激情性bbbbbb| kizo精华| 亚洲精品,欧美精品| 校园人妻丝袜中文字幕| 成人亚洲精品av一区二区| 网址你懂的国产日韩在线| 亚洲综合色惰| 18禁裸乳无遮挡免费网站照片| 别揉我奶头 嗯啊视频| 亚洲在久久综合| 少妇丰满av| 成人无遮挡网站| 亚洲天堂国产精品一区在线| 国产免费福利视频在线观看| 一级片'在线观看视频| 国产淫片久久久久久久久| 男女无遮挡免费网站观看| 91久久精品电影网| 国产精品三级大全| 51国产日韩欧美| 欧美变态另类bdsm刘玥| 久久精品人妻少妇| 国产高清国产精品国产三级 | 免费观看无遮挡的男女| 亚洲精品乱码久久久v下载方式| 久久久久久久久久久免费av| 草草在线视频免费看| 91久久精品国产一区二区成人| 在线亚洲精品国产二区图片欧美 | 免费黄频网站在线观看国产| 身体一侧抽搐| 九色成人免费人妻av| 欧美3d第一页| 久久久精品免费免费高清| 婷婷色麻豆天堂久久| 少妇裸体淫交视频免费看高清| 免费看日本二区| 久久亚洲国产成人精品v| 麻豆久久精品国产亚洲av| 一级a做视频免费观看| 51国产日韩欧美| 麻豆成人av视频| 欧美日韩在线观看h| 毛片一级片免费看久久久久| 国产乱人偷精品视频| 伦理电影大哥的女人| 一本色道久久久久久精品综合| 久久国内精品自在自线图片| 国产探花在线观看一区二区| 亚洲精品乱码久久久久久按摩| 18禁在线播放成人免费| 嘟嘟电影网在线观看| 亚洲经典国产精华液单| 一区二区av电影网| 禁无遮挡网站| 少妇的逼水好多| 久久国产乱子免费精品| 欧美成人a在线观看| 中文字幕久久专区| 日韩人妻高清精品专区| av在线天堂中文字幕| 中文字幕制服av| 熟女电影av网| 国产亚洲午夜精品一区二区久久 | 1000部很黄的大片| 国产精品精品国产色婷婷| 亚洲国产成人一精品久久久| 国产老妇女一区| 国产一区二区亚洲精品在线观看| 青春草国产在线视频| 人妻制服诱惑在线中文字幕| 国产精品一及| 久久久久国产精品人妻一区二区| 亚洲性久久影院| 亚洲伊人久久精品综合| 亚洲av免费高清在线观看| 欧美变态另类bdsm刘玥| 97人妻精品一区二区三区麻豆| 一本久久精品| 97超碰精品成人国产| 日本午夜av视频| 2021少妇久久久久久久久久久| 免费av观看视频| 国产一区二区在线观看日韩| 一本久久精品| 99久久精品一区二区三区| 只有这里有精品99| 卡戴珊不雅视频在线播放| 国产日韩欧美在线精品| 2022亚洲国产成人精品| 一区二区三区免费毛片| 久热这里只有精品99| 婷婷色综合大香蕉| 最近手机中文字幕大全| 国产精品国产av在线观看| 亚洲精品成人久久久久久| 美女cb高潮喷水在线观看| 一级毛片aaaaaa免费看小| 日日啪夜夜撸| 亚洲精品久久午夜乱码| 久久久久久久国产电影| 欧美变态另类bdsm刘玥| 少妇被粗大猛烈的视频| 欧美国产精品一级二级三级 | 免费大片黄手机在线观看| 亚洲av免费在线观看| 久久精品夜色国产| 亚洲成色77777| 久久久久久国产a免费观看| 婷婷色av中文字幕| 国产高潮美女av| av国产精品久久久久影院| 日韩电影二区| 男女下面进入的视频免费午夜| 国内精品宾馆在线| 亚洲无线观看免费| 色吧在线观看| 麻豆精品久久久久久蜜桃| 美女脱内裤让男人舔精品视频| 日韩强制内射视频| 又爽又黄a免费视频| 一区二区三区精品91| 欧美成人一区二区免费高清观看| 色5月婷婷丁香| av免费在线看不卡| 成年免费大片在线观看| 天堂中文最新版在线下载 | 少妇人妻精品综合一区二区| 五月开心婷婷网| 久久久国产一区二区| 亚洲av不卡在线观看| 欧美精品人与动牲交sv欧美| 国产免费视频播放在线视频| 国产午夜精品一二区理论片| 亚洲精品成人久久久久久| 乱系列少妇在线播放| 三级经典国产精品| 狂野欧美激情性xxxx在线观看| 国产综合懂色| 婷婷色综合大香蕉| 少妇 在线观看| 亚洲成色77777| 综合色丁香网| 国产精品麻豆人妻色哟哟久久| 国产 一区精品| 免费观看的影片在线观看| 伊人久久精品亚洲午夜| 一级a做视频免费观看| 蜜桃久久精品国产亚洲av| 日日啪夜夜爽| 日日啪夜夜撸| 美女cb高潮喷水在线观看| 久久精品国产亚洲av涩爱| 亚洲精品色激情综合| 嘟嘟电影网在线观看| 日本免费在线观看一区| 啦啦啦中文免费视频观看日本| 国产伦精品一区二区三区四那| 日本wwww免费看| 美女高潮的动态| 国内精品美女久久久久久| 欧美成人一区二区免费高清观看| 国产精品人妻久久久久久| 精品午夜福利在线看| 欧美极品一区二区三区四区| 神马国产精品三级电影在线观看| 国产免费一区二区三区四区乱码| 久久亚洲国产成人精品v| 午夜激情久久久久久久| 久久精品国产亚洲av天美| 亚洲在线观看片| 亚洲自拍偷在线| 亚洲成人久久爱视频| 在线天堂最新版资源| 嫩草影院新地址| 欧美老熟妇乱子伦牲交| 亚洲国产欧美在线一区| 亚洲精品一区蜜桃| 国产真实伦视频高清在线观看| 欧美xxxx性猛交bbbb| 国产免费一级a男人的天堂| 性插视频无遮挡在线免费观看| 国产成人freesex在线| 激情 狠狠 欧美| 国产成人精品一,二区| 国产成人免费观看mmmm| 久久精品国产亚洲网站| 久久久久久九九精品二区国产| 亚洲精品国产av成人精品| 亚洲色图av天堂| 超碰av人人做人人爽久久| 青春草视频在线免费观看| 99久久精品热视频| 国产v大片淫在线免费观看| 免费高清在线观看视频在线观看| 在现免费观看毛片| 国产精品一区二区在线观看99| 午夜亚洲福利在线播放| 成人午夜精彩视频在线观看| 日韩制服骚丝袜av| 香蕉精品网在线| 久久99蜜桃精品久久| 狂野欧美激情性bbbbbb| 大话2 男鬼变身卡| 精品久久久精品久久久| 精品少妇黑人巨大在线播放| 久久久精品欧美日韩精品| 久久久精品免费免费高清| 精品午夜福利在线看| 极品少妇高潮喷水抽搐| 日本wwww免费看| 如何舔出高潮| 自拍偷自拍亚洲精品老妇| 久久人人爽人人片av| 男的添女的下面高潮视频| 最近手机中文字幕大全| 国产老妇女一区| 久久精品久久精品一区二区三区| 午夜免费观看性视频| 大码成人一级视频| 真实男女啪啪啪动态图| 亚洲国产精品专区欧美| 内地一区二区视频在线| eeuss影院久久| 性色av一级| 只有这里有精品99| 国产乱来视频区| 神马国产精品三级电影在线观看| 国产精品99久久99久久久不卡 | 久久久久性生活片| 久久精品人妻少妇| 亚洲欧美中文字幕日韩二区| 欧美三级亚洲精品| 春色校园在线视频观看| 全区人妻精品视频| 男男h啪啪无遮挡| 少妇高潮的动态图| 国产黄片视频在线免费观看| 日韩国内少妇激情av| 精品少妇久久久久久888优播| 特级一级黄色大片| 在线观看美女被高潮喷水网站| 国产毛片在线视频| 深爱激情五月婷婷| 色5月婷婷丁香| 久久综合国产亚洲精品| 免费看不卡的av| 激情 狠狠 欧美| 亚洲国产精品999| 国产综合精华液| 日韩国内少妇激情av| 日本av手机在线免费观看| 人人妻人人爽人人添夜夜欢视频 | 成年免费大片在线观看| 久久99精品国语久久久| 国产精品一二三区在线看| 男人舔奶头视频| 97精品久久久久久久久久精品| 亚洲精品日韩在线中文字幕| 黄色视频在线播放观看不卡| 日韩人妻高清精品专区| 最近2019中文字幕mv第一页| 两个人的视频大全免费| 下体分泌物呈黄色| 欧美精品一区二区大全| 51国产日韩欧美| 国产黄色免费在线视频| 热re99久久精品国产66热6| 国产真实伦视频高清在线观看| 超碰97精品在线观看| 亚洲精品成人久久久久久| 亚洲国产成人一精品久久久| 欧美日韩精品成人综合77777| 嘟嘟电影网在线观看| 3wmmmm亚洲av在线观看| 国产爽快片一区二区三区| 亚洲国产av新网站| videos熟女内射| 国产精品人妻久久久影院| 精品视频人人做人人爽| 欧美少妇被猛烈插入视频| 亚洲欧美成人精品一区二区| 青青草视频在线视频观看| 亚洲最大成人中文| 毛片一级片免费看久久久久| 大陆偷拍与自拍| 最近手机中文字幕大全| 国内精品美女久久久久久| 色婷婷久久久亚洲欧美| 亚洲精品aⅴ在线观看| 51国产日韩欧美| 日本与韩国留学比较| 欧美成人精品欧美一级黄| 一区二区av电影网| 国产亚洲av片在线观看秒播厂| 在线观看三级黄色|