• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    RNAGCN:RNA tertiary structure assessment with a graph convolutional network

    2022-11-21 09:32:26ChengweiDeng鄧成偉YunxinTang唐蘊芯JianZhang張建
    Chinese Physics B 2022年11期
    關(guān)鍵詞:王駿張建

    Chengwei Deng(鄧成偉) Yunxin Tang(唐蘊芯) Jian Zhang(張建)

    Wenfei Li(李文飛)1,2, Jun Wang(王駿)1,2, and Wei Wang(王煒)1,2,?

    1Collaborative Innovation Center of Advanced Microstructures,School of Physics,Nanjing University,Nanjing 210008,China

    2Institute for Brain Sciences,Nanjing University,Nanjing 210008,China

    RNAs play crucial and versatile roles in cellular biochemical reactions.Since experimental approaches of determining their three-dimensional (3D) structures are costly and less efficient, it is greatly advantageous to develop computational methods to predict RNA 3D structures. For these methods, designing a model or scoring function for structure quality assessment is an essential step but this step poses challenges. In this study, we designed and trained a deep learning model to tackle this problem. The model was based on a graph convolutional network(GCN)and named RNAGCN.The model provided a natural way of representing RNA structures, avoided complex algorithms to preserve atomic rotational equivalence,and was capable of extracting features automatically out of structural patterns. Testing results on two datasets convincingly demonstrated that RNAGCN performs similarly to or better than four leading scoring functions.Our approach provides an alternative way of RNA tertiary structure assessment and may facilitate RNA structure predictions. RNAGCN can be downloaded from https://gitee.com/dcw-RNAGCN/rnagcn.

    Keywords: RNA structure predictions, scoring function, graph convolutional network, deep learning, RNApuzzles

    1. Introduction

    RNAs play crucial and versatile roles in cellular biochemical reactions, such as encoding, decoding, catalysis,[1]gene regulations,[2]and others. These functions are closely related to the three-dimensional (3D) structures of the RNAs.To determine RNA 3D structure, experimental technics, including cryo-electron microscopy, x-ray crystallography, and nuclear magnetic resonance (NMR) spectroscopy are usually employed. Since these technics are costly and inefficient,computational methods have been developed to predict RNA structures.[3–19]Computational algorithms usually include two steps: (i) generating structural candidates and (ii) selecting structures most likely to be the native. The second step requires a good scoring function that can assess the quality of the candidates.

    Traditional scoring functions[20,21]are physics-based or knowledge-based, to name a few, Rosetta,[4,22]3dRNAscore,[23]RASP,[24]RNA KB potential,[25]DFIRERNA,[26]and rsRNASP.[27]Among them, 3dRNAscore is an all-atom statistical potential that combines distance-dependent and dihedral-dependent energies;[23]it is more efficient than other potentials in recognizing and ranking native state from a pool of near-native decoys. The rsRNASP potential is composed of short- and long-ranged energies, distinguished by residue separation along sequence;[27]extensive tests showed that it has higher or comparable performance against other leading potentials, dependent on specific testing datasets. In recent years, machine learning approaches achieved great success in many fields, including computer vision, natural language modelling,[28,29]medical diagnosis,[30]physics,chemistry, computational biology, and so on.[31–34]Inspired by these successes, our group developed a scoring function for the assessment of RNA tertiary structure based on a three-dimensional convolutional network and named it RNA3DCNN.[35]

    Applications of graph convolutional network(GCN)[36–41]to represent molecular structures have been quite successful.[42–47]For example, Foutet al.[43]modelled proteins as graphs at residue level and then predicted protein interfaces. Federicoet al.[44]introduced GCNs (GraphQA)to tackle the problem of quality assessment of protein structures, and showed that using only a few features will result in state-of-the-art results. Soumyaet al.[45]developed ProteinGCN for protein structures assessment; they built graphs to model spatial and chemical relations between pairs of atom.Zheet al.[46]built a GCN network, named GraphCPI, which aggregated chemical context of protein sequences and structural information of compounds to tackle the problem of compound–protein interaction. Huanget al.[47]introduced a model called GCLMI based on GCN and an auto-encoder to predict IncRNA–miRNA interactions.

    Inspired by the notion that a graph provides a more natural representation of 3D RNA structures and hence may bring better assessing performance, in this study, we upgrade our previous model by changing the convolutional network to a GCN network. Specifically, we developed a network model based on GCN, named RNA graph convolutional network(RNAGCN),to perform quality assessment of RNA structures.We trained and tested the model and compared the results with several leading scoring functions.

    2. Materials and methods

    In this section,we first introduce the input and output of RNAGCN, and then the architecture of the model. Next, we present the datasets used for training and evaluation,followed by a description of the metrics used to evaluate the performance of the model. At last,we describe the loss function and the training procedures.

    2.1. Graph representation of RNA structures

    RNA structures can be naturally represented as graphs,with nodes modeling atoms and edges modeling their relative spatial position in three-dimensional space. Since RNA sizes vary, we split an input RNA into many “l(fā)ocal environments”and converted each into a graph. Specifically, for theithnucleotide along the sequence, we defined it as the central nucleotide of its “l(fā)ocal environment” constructed by including all its nucleotide neighbors. A nucleotide was defined to be a neighbor if any of its heavy atoms (other than a hydrogen atom) was within a threshold spatial distance of any heavy atoms of the central nucleotide. The distance threshold was set to be 14 ?A.This is to construct a large enough local environment for the central nucleotide.With this design,the model can be scaled to handle RNAs of arbitrary size. Thus, for an RNA of lengthNs,we constructedNslocal environments represented byNsgraphs.

    In general,a graph is defined as a set

    whereVdenotes the set of nodes andEdenotes the set of edges,respectively.X ∈R|V|×NandU ∈R|?|×Eare the feature matrices ofVandE,respectively.N,Eare the number of nodes and number of edges,respectively.

    All heavy (non-hydrogen) atoms in a local environment were treated as nodes. A node was represented as a one-hot vector of lengthNt,whereNtis the total number of atom types,which is 54,based on AMBER99SB force field.

    Edges were defined between neighboring atoms. Each atom was connected by an edge with itsKnearest atoms in space, whereKwas set to be 14. Edges were directed and each included five features. The first was the spatial distance betweenSandT, which was one of its neighbors. The 2nd,3rd,and 4th were direction features,which were the three projections of the unit vector pointing fromStoTin the internal coordinate system centered onS,i.e.,

    whereAdenotes the ‘C1’ atom andBthe ‘C5’ atom in a nucleotide. The fifth feature had a value of 1 or 0,depending on whether there is a chemical bond betweenSandT.

    2.2. Output of RNAGCN:quality score

    The graphs representing the local nucleotide environment constructed for each nucleotide from its neighbors were used as input to the GCN model. The model output was a scalar score indicating the quality of the input. During training,the output score measured the difference of local environment from the ground truth, i.e., the RMSD of the input structure with respect to the experimental one. During the inferring operation,the network output was the predicted score indicating the RMSD of the input versus the experimental structure that in this operation was an unknown to the network. The scores thus obtained for theNsinput environments of an RNA were averaged to get the final score,which reflected the overall difference of the input structure to the experimental one, with a lower score corresponding to a higher input quality (a better approximation of the unknown experimental structure).

    2.3. Architecture of RNAGCN

    RNAGCN is a deep learning model based on GCN that is used to extract features from 3D structures of RNA. The model architecture is shown in Fig. 1. The input to the model is an RNA local graph,converted from an RNA local environment. Below the input layer,there are five serially connected graph convolution layers that operate on the graph sequentially.Residual modules and skip connections were adopted to solve the gradient vanishing problem often seen in deep neural networks.[38,40]Serially connected below the GCN layers,there is a convolution layer with a 1×1 kernel, followed by a graph global max pooling layer, and then the final layer of a fully-connected network that outputs a score as prediction.The model contains about 378k parameters in total.

    In the five GCN layers, graph convolution operations were used to update the nodes’representations iteratively. The details of the operations are shown on the right side of Fig.1.

    Fig.1. Overview of the RNAGCN model. The input is an RNA local graph. The network contains five graph convolution layers,a convolution layer with a 1×1 kernel,a graph global max pooling layer,and a fully-connected network that outputs a score as prediction. The right panel presents details of feature operations using graph convolution with residual connections.

    Following the GCN layers, a convolution layer with a 1×1 kernel was used to perform the aggregation of information fromLGCN nodes with different weights. These weights were learned from data during training. As a result,Xwas reduced fromRL×|V|×NtoR|V|×N.

    A global max pool layer was used to mix nodes representations. Max values ofX ∈R|V|×Nalong|V|were computed one by one. As a result,Xwas further reduced,fromR|V|×NtoRN.

    Finally,a fully-connected network was used to reduceXfromRNtoR1.It included a layer of size 64 with ReLU activation function,and a layer of size 1 without activation function.

    Notably, the representation of edges was kept static during graph convolution,according to the generally accepted notion that updating nodes’features alone is usually sufficient to achieve reasonable model performance.

    2.4. Datasets

    We built our datasets based on the non-redundant set of RNA 3D Hub[48](http://rna.bgsu.edu/rna3dhub/nrlist,Release 3.102, 2019-11-27). RNAs forming complex with proteins or other molecules were removed; only pure RNA structures were kept. Short RNAs with length smaller than eight nucleotides were also removed. The remaining 610 RNAs were split into training, validation and testing datasets. The infernal program[49]was used to ensure that no RNA in the testing dataset belonged to the same RFAM family as in the training and validation datasets. The resulting three datasets contained 426, 92, and 92 RNAs, respectively. Hereafter, the testing dataset of 92 RNAs is referred to as Test-I.

    RNA decoys are needed to train and test scoring functions. From the experimental structure of each RNA, we generated the corresponding decoys with high-temperature molecular dynamics (MD) simulations, carried out using Gromacs.[50]In brief, each RNA molecule was solved in a TIP3P water box,and metal ions were added to neutralize the system. The system was then subjected sequentially to energy minimization, NVT equilibration, and NPT equilibration at 300 K.Finally,high temperature MD was used to denature the structure by gradually increasing the temperature from 300 K to 660 K.The force field was AMBER99SB.Each simulation lasted for 10 ns. For each trajectory,we calculated the RMSD of each frame with respect to the corresponding experimental structure and randomly selected 200 decoys in the RMSD range from 0 to 20 ?A. The decoys, together with all the experimental structures,constructed the training,validation,and testing datasets,as summarized in Table 1.

    Fig.2. Distributions of(a)RMSD and(b)length for four datasets. Note that RNA 6qkl in the training set is not shown,for its length is too long(1158 nucleotides).

    We prepared another testing dataset based on the RNApuzzles-standardized dataset,[51]which included 22 RNAs.Again, for each RNA in this set, the same denaturing procedure was carried out by high-temperature MD to generate decoys, and 200 decoys were randomly selected. The resulting testing dataset,referred to as Test-II hereafter,contained 4422 RNA structures.

    The distribution of RMSDs and lengths for the abovementioned datasets are given in Fig.2.

    Table 1. Information of datasets.

    2.5. Evaluation metrics

    Enrichment score (ES) and Pearson correlation coefficient(PCC)were used to evaluate the performance of scoring function.

    Enrichment score(ES)[23,25–27,35,52]was defined as

    whereStop10%andRtop10%indicated the best-scored 10% decoys and the 10% with the lowest RMSDs, respectively.Ndecoyswas the number of decoys associated with the RNA of concern. ES values ranged from 0 to 10, with larger value indicating better performance.

    The Pearson correlation coefficient (PCC)[24,27,52]indicates the magnitude of linear correlation between two variables and is defined as

    whereSiandRiare the scores given by the scoring function and the RMSD of the structure, respectively. PCC assumes a value between 0 and 1.

    2.6. Loss function

    The loss function, used as the optimizing target of the model,was defined as

    whereRiandSiwere RMSD and the predicted score of theithlocal environment,respectively.Nenvswas the total number of local environments in the whole training dataset. In this study,it amounts to about 107.

    2.7. Training strategies

    All the RNAs were partitioned into local environments,which were then randomly shuffled and fed to the network in batches. The batch size per GPU was set to 64 and the total batch size was 384, since 6 GPUs were used in parallel computation. Python modules Deep Graph Library and Pytorch.Distributed were used for model construction and computation parallelization,respectively. Mini-batch gradient descent algorithm was employed to train the model.Specifically,the Adam optimizer with parametersβ1=0.9,β2=0.98 was used. The initial learning rate was 0.004 and it would be decreased by half if the training loss no longer dropped for 4 epochs.

    3. Results

    In this section,we present the results of our model experiments and compare the performance of our model with four statistical potentials, including RASP, Rosetta, 3dRNAscore,and rsRNASP, which are currently the most popular in the field.

    3.1. Performance on Test-I

    Table 2 shows the performance of the five models in assessing the quality of structures in the Test-I dataset, which contained 92 RNAs and 200 decoys associated with each.We used two criteria, labeled as Top-1 and Top-5, to reflect whether the experimental RNA was ranked the first,or ranked among the best five, respectively. As judged by Top-1, 3dRNAscore and rsRNASP performed best;they identified 91 and 86 out of 92 RNAs,respectively. RNAGCN(our model)performed as a third (79/92). Judged by the criterion Top-5, all models performed similarly. According to comparisons based on these two criteria, our model is the third best among all.This outcome is tentatively attributed to the strict criteria we used to construct the testing set: we made sure there are no overlapping of RFAM families between the testing set and the training datasets.

    Table 2. Performance of five models on Test-I.

    Fig.3.(a)Experimental structure of RNA 5dcv.(b)Score-RMSD plots,showing the correlation between the scores predicted by the model and the RMSDs of the structures. The purple crosses mark the experimental structures, orange crosses indicate the structures that were scored better than the experimental ones, and red crosses mark the best-scored structures if they were not the experimental one.

    Table 2 also presents the average values for ES and PCC that indicate the strength of correlations between the ground truth and prediction.Our model is superior to the others by the ES measure, close second best behind RASP by the average PCC, and clearly superior by the measure sensitive to nearnative structures (PCC with RMSD<4 ?A). The latter result indicates that our model excels at discriminating small structural changes with respected to the native one.

    In Fig. 3(b) we present the score-RMSD plots for each of the five models, computed using RNA-5dcvas an example(Fig.3(a)),which is a 95-nucleotid-long fragment fromP.horikoshiiRNase.[53]The data show that RASP and Rosetta failed to identify the experimental structure as the best one.In contrast,3dRNAscore,rsRNASP and our model ranked the experimental one as the best structure(as indicated by overlapping symbols for best score and native). Moreover,our model appeared to show a better score-RMSD correlation, particularly at small RMSD ranges.For a further proof,we calculated the average PCC values for RNAs within the 0 to 4 ?A range of RMSD and found them to be 0.77,0.47,-0.65,0.61,and 0.70 for RNAGCN, RASP, Rosetta, 3dRNAscore, and rsRNASP,respectively. By this measure, our model clearly exhibited a significantly higher correlation between scores and RMSDs.This result is also consistent with that bottom line in Table 2,which shows that our model gives the best correlation in the small RMSD range. Moreover, for this specific RNA,the ES values of RNAGCN, RASP, Rosetta, 3dRNAscore, and rsRNASP are 5.0,3.0,1.0,0.5,and 5.0,respectively,also consistent with the ES result presented in Table 1.

    The ES, PCC, and score-RMSD plots for each RNA in Test-I are listed in the supplementary material.

    3.2. Performance on Test-II

    We also analyzed the performance of the five models on the Test-II testing dataset, which is composed of 22 RNAs taken from RNA-puzzles-standardized datasets. The result is summarized in Table 3.

    Table 3 shows that 3dRNAscore identified all 22 experimental structures correctly and our model identified just one less, 21 out of 22 RNAs, a close second. Judged by the ES values, our model is 6.5, which is very closed to the best 6.7(rsRNASP).The average PCC of our model is 0.88,also very close second to the best value 0.90 (rsRNASP). Within the small RMSD range (<4 ?A), the PCC value of our model is again close to the best 0.76 (rsRNASP). Detailed results for each RNA in Test-II can be found in the supplementary material.

    Table 3. Performance of five models on Test-II.

    Fig. 4. (a) Experimental structure of puzzle-2 in the RNA-puzzles dataset. (b) Score-RMSD plots, showing the correlation between the predicted scores and the RMSDs of the structure. The purple, orange,and red crosses are defined in the same way as in Fig.3.

    Figure 4(b)shows the score-RMSD plots for puzzle-2 in Test-II.The pattern of performances is similar to that of Test-I in Fig.3. Three models, including 3dRNAscore, rsRNASP,and ours,correctly identified the experimental structure out of decoys. For this specific RNA, the ES value of our model is 9.5,very closed to perfect value 10;and the PCC value is 0.95.Our model and RASP appear to show a good correlation between scores and RMSDs, while Rosetta, 3dRNAscore, and rsRNASP feature a big gap in scores between the experimental structure and the decoys. The score-RMSD plots for other RNAs in Test-II are generally similar to Fig.4 and listed in full in the supplementary material.

    We compiled a detailed ranking of the five models for Test-II and presented the results as a colormap in Fig.5. For each RNA in Test-II and each scoring function, we selected the best-scoredNb(a pre-set variable, “b” indicates best) decoys and computed their average RMSD. We then ranked five scoring functions based on these RMSDs and represented the ranks with different colors, with lighter colors indicating higher ranks. It can be seen that RNAGCN has the largest area of light colors(light yellow and gold),indicating that our model gives the lowest RMSDs more often than the others.

    Following the usual practice in the field of machine learning,we plotted receiver operating characteristic(ROC)curves on difference choices ofNb, used as a variable threshold to select structures as the best predicted ones. For example, ifNbwas 10, we trusted the best scored 10 structures ranked by the model as the predicted native structures. To test the model performance under different choices ofNb,we first defined a native ensemble as all structures with RMSD smaller than 2 ?A,and then defined all structures within the native ensemble as positive samples and the others as negative. These structures and the associated binary classification labels were used to construct the ground-truth dataset, following conventions in machine learning field.During predictions,the bestNbstructures given by model were treated as the predicted positives,while the leftover was treated as the predicted negatives.Under varying choices ofNb, we calculated the true-positive rate (TPR) and false-positive rate (FPR) and plotted their relationship as ROC curves in Fig. 6. It can be seen that the ROC curve of our model almost envelopes (i.e., maximizes)the other four curves. Quantitatively, the areas under ROC(AUC) of our model for Test-I and Test-II are 0.97 and 0.89,respectively; both represent the best performance among the all scoring functions.

    Fig. 5. Ranking of five models based on the average RMSDs on the best-scored Nb structures, where Nb is a variable plotted on the x-axis.Each row corresponds to one RNA in Test-II(a total of 22,note their labels are not continuous in the standard dataset). The ranks are indicated by the color bar on the right;the lighter the color,the higher the rank.

    Fig.6. ROC curves of five scoring functions on different choices of threshold Nb. Panels (a) and (b) are results obtained for the datasets Test-I and Test-II,respectively.

    4. Discussion and conclusions

    Applications of deep learning technology in molecular representation learning have achieved impressive progress in recent years. Inspired by previous works,we explored the application of graph convolutional network in RNA 3D structure assessment tasks. We compiled two datasets with MD simulations and trained a GCN neural network. We tested the model and evaluated its performance using four leading scoring functions. The testing dataset Test-I contained 92 RNAs. Dataset Test-II contained 22 RNAs,a subset of the RNA-puzzles standardized dataset.

    For both testing datasets,the ability of our GCN model to identify experimental structures closely approached the best–3dRNAscore on Test-I and rsRNASP on Test-II,respectively.The ES metric,measuring the overlap of the near-native structures and the best scored structures, indicated that our model is the best one for distinguish near-native structures (<4 ?A)on Test-I while the second one on Test-II. Additionally, we ranked five models based on average RMSD of the best scored structures, and compared their ROC curves and AUCs; both experiments ranked our model as superior to the others.

    It is interesting to compare our machine learning model with the statistical potentials based on inverse-Boltzmann equation,particularly,3dRNAscore and rsRNASP.According to the presented results, our model seems to perform slightly worse in identifying native structures (Tables 2 and 3) and slightly better in comparison of AUCs. This outcome may be tentatively attributed to the strict criteria we used to construct the testing set: we made sure there are no overlapping of RFAM families between the testing set and the training datasets. Besides, we did not train the model over and over to increase the metrics,since too much training might lead to over-fitting and then decrease generalization ability, even the training and testing datasets were strictly separated.

    In general, the advantages of statistical potentials based on inverse-Boltzmann equation are their clear physical picture,good generalization ability to unseen structures,and fast computation speed. In contrast,machine learning based potentials are black-boxes and have unclear generalization ability and arguably slow speed.However,the advantages of machine learning approaches are also prominent. First,it needs not years of studies looking for proper energy forms and choice of reference states.[52]For RNA structures,it has taken people a dozen years to achieve the current performance. While a similar performance can be achieved by simply taking a general deep network(with slightly modifying the input and output layers)and training it properly. This advantage becomes more prominent if people need to migrate model to new molecules. Second,deep neural networks have found physical patterns similar to those found by human,and have the potential to find new patterns not-known yet. Machine learning based scoring models deserve to be studied.

    A recently introduced geometric deep learning approach named ARES[54]showed good performance at scoring RNA structures, but unfortunately, it is not yet possible to directly compare it to ours. The authors of ARES have not released their trained model, and we failed to reproduce their results by attempting to rebuild the model from downloaded codes and training the same network with their datasets. Moreover,ARES was trained on datasets generated with FARFAR2,different from the datasets we generated from MD simulations.A fair comparison of these two models requires identical data or data sampled from independently identical distribution for training and testing.

    Our model has several advantages. First, it is natural to represent the geometric and topologic information about RNA tertiary structures as a graph. Second, because a graph automatically provides geometric invariant representation regarding atomic translation and rotation, there is no need to design complex convolution operations such as those used in ARES.Third,because our model can directly learn from spatial patterns of atoms on its own, it requires, unlike physicsbased scoring functions,no prior physical knowledge. Therefore, our approach can be easily extended to other molecular systems with little modification. Forth, the design of splitting structures into local environments makes the model scalable,enabling us to treat RNAs of arbitrary size. At last,this study showed that our model performed well particularly for near-native structures. We tentatively attributed this feature to the graph representation of tertiary structures, for both graph topology and edge features are sensitive to the changes in positions of atoms.

    There are two notable limitations of the present work.One limit is computational, stemming from the huge memory consumption of graphs and the massive computations involved in graph convolution operations. As both factors limit the size of the graph,and the number of neighbor atoms around the central nucleotide,they may raise problems handling very long-range interactions that are relevant. The second is algorithmic,as in the current version of the model only node features in the network were updated while edge features were kept static. Presumably,updating both in an upgraded version may slightly improve model performance.

    Acknowledgements

    This study was funded by the National Natural Science Foundation of China(Grant Nos.11774158 to JZ,11934008 to WW,and 11974173 to WFL).The authors acknowledge High Performance Computing Center of Advanced Microstructures,Nanjing University for the computational support.

    猜你喜歡
    王駿張建
    護苗
    中國火炬(2024年4期)2024-04-17 11:27:44
    趕海記憶
    鷸蚌相爭,漁人得利
    Diffusion of nucleotide excision repair protein XPA along DNA by coarse-grained molecular simulations?
    豆腐里的愛
    本期焦點人物:張建一
    綠色中國(2019年13期)2019-11-26 07:10:50
    《CT成像:基本原理、偽影與誤區(qū)》已出版
    書寫精氣神
    榮譽雜志(2015年2期)2015-07-08 14:23:56
    肩關(guān)節(jié)生物力學(xué)
    真情假意
    故事林(2007年12期)2007-05-14 15:37:51
    91精品伊人久久大香线蕉| 国产男女超爽视频在线观看| 欧美 日韩 精品 国产| 亚洲精品456在线播放app| 亚洲四区av| 午夜福利视频1000在线观看| 日本欧美国产在线视频| 亚洲欧美日韩卡通动漫| 青春草亚洲视频在线观看| 在线观看美女被高潮喷水网站| 久久99热这里只频精品6学生| 一级毛片我不卡| 韩国高清视频一区二区三区| 国产大屁股一区二区在线视频| www.av在线官网国产| 深夜a级毛片| 国产成人福利小说| 男人爽女人下面视频在线观看| 国产大屁股一区二区在线视频| 欧美最新免费一区二区三区| 亚洲一级一片aⅴ在线观看| 视频中文字幕在线观看| 极品教师在线视频| 成人免费观看视频高清| 99热网站在线观看| 国产免费一区二区三区四区乱码| 高清在线视频一区二区三区| 欧美三级亚洲精品| 99热国产这里只有精品6| 少妇裸体淫交视频免费看高清| 99热6这里只有精品| 成年女人在线观看亚洲视频 | 亚洲国产欧美人成| 97精品久久久久久久久久精品| 久久精品熟女亚洲av麻豆精品| 欧美xxⅹ黑人| 2018国产大陆天天弄谢| 狂野欧美激情性bbbbbb| 色5月婷婷丁香| 亚洲av欧美aⅴ国产| 日本色播在线视频| 亚洲欧美成人综合另类久久久| 国产av不卡久久| 综合色av麻豆| 色视频在线一区二区三区| 一级毛片我不卡| av女优亚洲男人天堂| 国产精品一区www在线观看| 亚洲四区av| 一级黄片播放器| 中文字幕久久专区| 免费看av在线观看网站| 成人美女网站在线观看视频| 内射极品少妇av片p| 国产精品一区二区性色av| 亚洲美女视频黄频| 青春草亚洲视频在线观看| 亚洲精品国产av成人精品| 熟女人妻精品中文字幕| 国产v大片淫在线免费观看| 国产成人精品一,二区| 夫妻性生交免费视频一级片| 亚洲自偷自拍三级| 狂野欧美激情性bbbbbb| 不卡视频在线观看欧美| 久久99热这里只有精品18| 嫩草影院入口| 99热这里只有精品一区| 欧美区成人在线视频| 亚洲国产av新网站| 日韩免费高清中文字幕av| 亚洲久久久久久中文字幕| 一级爰片在线观看| 中文资源天堂在线| 国产毛片a区久久久久| 高清午夜精品一区二区三区| 亚洲精品aⅴ在线观看| 3wmmmm亚洲av在线观看| 久久久国产一区二区| 亚洲色图综合在线观看| 少妇熟女欧美另类| 国产精品麻豆人妻色哟哟久久| 免费av观看视频| 不卡视频在线观看欧美| 国产女主播在线喷水免费视频网站| 国内少妇人妻偷人精品xxx网站| 精品国产三级普通话版| 91久久精品国产一区二区三区| 听说在线观看完整版免费高清| 亚洲精品国产av成人精品| 99热这里只有是精品在线观看| 国产日韩欧美在线精品| 成人高潮视频无遮挡免费网站| 人妻夜夜爽99麻豆av| 精品久久久久久电影网| 久久ye,这里只有精品| 欧美xxxx黑人xx丫x性爽| 午夜爱爱视频在线播放| 中文字幕免费在线视频6| 久久精品久久久久久久性| 亚洲伊人久久精品综合| 日日撸夜夜添| 美女xxoo啪啪120秒动态图| 在现免费观看毛片| 丝瓜视频免费看黄片| 在线播放无遮挡| 免费看a级黄色片| 嫩草影院新地址| 国产在线一区二区三区精| 在线天堂最新版资源| 自拍欧美九色日韩亚洲蝌蚪91 | 亚洲精品国产成人久久av| 麻豆乱淫一区二区| 国产精品一区二区在线观看99| 日韩不卡一区二区三区视频在线| 亚洲综合精品二区| 亚洲精华国产精华液的使用体验| 久久99蜜桃精品久久| 男女下面进入的视频免费午夜| 少妇高潮的动态图| 午夜福利网站1000一区二区三区| 亚洲天堂av无毛| 有码 亚洲区| 国产av国产精品国产| 午夜福利视频1000在线观看| 韩国高清视频一区二区三区| 国产色爽女视频免费观看| 91精品一卡2卡3卡4卡| 亚洲成人av在线免费| 亚洲第一区二区三区不卡| 2022亚洲国产成人精品| 大码成人一级视频| 亚洲高清免费不卡视频| 久久久久国产精品人妻一区二区| a级一级毛片免费在线观看| av在线观看视频网站免费| 日本爱情动作片www.在线观看| 国产在线一区二区三区精| 麻豆久久精品国产亚洲av| 国产亚洲最大av| 国产一级毛片在线| 夫妻性生交免费视频一级片| 噜噜噜噜噜久久久久久91| 久热久热在线精品观看| 永久免费av网站大全| 中国国产av一级| 亚洲国产成人一精品久久久| 交换朋友夫妻互换小说| av在线天堂中文字幕| 久久久久久久国产电影| 国产乱人视频| 欧美成人一区二区免费高清观看| 18+在线观看网站| 欧美日韩精品成人综合77777| 最后的刺客免费高清国语| 可以在线观看毛片的网站| 街头女战士在线观看网站| 免费观看无遮挡的男女| 欧美变态另类bdsm刘玥| 亚洲av.av天堂| 国内精品宾馆在线| 国产精品国产av在线观看| 亚洲成人中文字幕在线播放| 人妻少妇偷人精品九色| 中文资源天堂在线| 亚洲av中文av极速乱| 国产一区亚洲一区在线观看| 能在线免费看毛片的网站| 18禁动态无遮挡网站| av专区在线播放| 国产成人精品久久久久久| av免费在线看不卡| 国模一区二区三区四区视频| 亚洲美女搞黄在线观看| 18禁在线播放成人免费| 亚洲天堂国产精品一区在线| 只有这里有精品99| 深爱激情五月婷婷| 日本黄色片子视频| 人人妻人人爽人人添夜夜欢视频 | 能在线免费看毛片的网站| 久久久精品免费免费高清| 亚洲精品成人久久久久久| 亚洲精品国产av成人精品| 国产精品人妻久久久久久| 国产精品av视频在线免费观看| 2021少妇久久久久久久久久久| 免费av观看视频| 九九久久精品国产亚洲av麻豆| 丝袜美腿在线中文| 午夜视频国产福利| 汤姆久久久久久久影院中文字幕| 99久久精品热视频| 女人十人毛片免费观看3o分钟| 如何舔出高潮| av在线亚洲专区| 国产成人精品久久久久久| av免费观看日本| 久久精品久久精品一区二区三区| 天堂俺去俺来也www色官网| 国产伦理片在线播放av一区| 神马国产精品三级电影在线观看| 国产精品精品国产色婷婷| 久久精品国产亚洲网站| 国产老妇女一区| 狂野欧美激情性xxxx在线观看| 男人狂女人下面高潮的视频| 亚洲最大成人中文| 美女主播在线视频| 啦啦啦中文免费视频观看日本| 欧美丝袜亚洲另类| 久久精品熟女亚洲av麻豆精品| 欧美日韩综合久久久久久| 最新中文字幕久久久久| 国产高清有码在线观看视频| 国产一级毛片在线| 免费大片黄手机在线观看| eeuss影院久久| 91精品伊人久久大香线蕉| 性色av一级| 亚洲精品日本国产第一区| 亚洲国产精品成人久久小说| 美女被艹到高潮喷水动态| 日韩欧美 国产精品| 国产一区亚洲一区在线观看| 男男h啪啪无遮挡| 晚上一个人看的免费电影| 国产伦理片在线播放av一区| 男女啪啪激烈高潮av片| 3wmmmm亚洲av在线观看| 最近中文字幕2019免费版| 街头女战士在线观看网站| 国产成人福利小说| 亚洲欧美清纯卡通| 精品亚洲乱码少妇综合久久| 最近中文字幕2019免费版| 丰满人妻一区二区三区视频av| 男人和女人高潮做爰伦理| 人妻 亚洲 视频| 亚洲综合色惰| 蜜桃久久精品国产亚洲av| 欧美精品一区二区大全| 最后的刺客免费高清国语| 99re6热这里在线精品视频| 亚洲av免费在线观看| 成人二区视频| 国产爱豆传媒在线观看| 最后的刺客免费高清国语| 内射极品少妇av片p| 日韩av免费高清视频| 成人亚洲精品av一区二区| 成人毛片a级毛片在线播放| 午夜亚洲福利在线播放| 国产精品久久久久久久久免| 久久久色成人| 热re99久久精品国产66热6| 91狼人影院| 久久热精品热| 爱豆传媒免费全集在线观看| 91在线精品国自产拍蜜月| 建设人人有责人人尽责人人享有的 | 久久国产乱子免费精品| 国产高清有码在线观看视频| 精品99又大又爽又粗少妇毛片| 欧美三级亚洲精品| 亚洲国产精品成人综合色| 亚洲丝袜综合中文字幕| 欧美日韩精品成人综合77777| 国产 精品1| 波野结衣二区三区在线| 免费观看性生交大片5| 欧美一级a爱片免费观看看| 五月伊人婷婷丁香| 建设人人有责人人尽责人人享有的 | 免费av观看视频| 日本wwww免费看| 国产视频内射| 国产乱人视频| 色婷婷久久久亚洲欧美| 亚洲成人av在线免费| 观看美女的网站| 欧美97在线视频| 中文字幕av成人在线电影| 久久久久久久午夜电影| 黄色视频在线播放观看不卡| 秋霞伦理黄片| 日韩一本色道免费dvd| 日韩大片免费观看网站| 好男人视频免费观看在线| av黄色大香蕉| 一级毛片黄色毛片免费观看视频| 黑人高潮一二区| 久久国内精品自在自线图片| 在线观看一区二区三区激情| 精品人妻视频免费看| 久久久久网色| av.在线天堂| 午夜免费鲁丝| 亚洲人成网站在线播| h日本视频在线播放| 99久久精品热视频| 大香蕉97超碰在线| 国产色爽女视频免费观看| 亚洲av日韩在线播放| 国产精品人妻久久久影院| 国产老妇伦熟女老妇高清| 日韩欧美一区视频在线观看 | 黄片wwwwww| 熟女人妻精品中文字幕| 永久网站在线| 一本一本综合久久| 在线a可以看的网站| 成人漫画全彩无遮挡| 大又大粗又爽又黄少妇毛片口| 中国三级夫妇交换| 亚洲av在线观看美女高潮| 欧美日韩在线观看h| 国产亚洲精品久久久com| h日本视频在线播放| 国产精品女同一区二区软件| 赤兔流量卡办理| 亚洲一级一片aⅴ在线观看| 特大巨黑吊av在线直播| av免费观看日本| 欧美3d第一页| 久久99精品国语久久久| 亚洲欧美日韩另类电影网站 | freevideosex欧美| 亚洲人与动物交配视频| 国产女主播在线喷水免费视频网站| 精品久久久久久久末码| 精品久久久噜噜| 亚洲精品视频女| 在线观看美女被高潮喷水网站| 亚洲精品456在线播放app| 一个人看的www免费观看视频| 一区二区av电影网| 插阴视频在线观看视频| 亚洲av成人精品一区久久| 久久精品人妻少妇| 少妇裸体淫交视频免费看高清| 内地一区二区视频在线| 中文天堂在线官网| 亚洲四区av| 精品国产露脸久久av麻豆| 国产成人a∨麻豆精品| 另类亚洲欧美激情| 欧美另类一区| 国产久久久一区二区三区| 国产免费一级a男人的天堂| 深爱激情五月婷婷| 中文字幕制服av| 国产精品久久久久久久电影| 免费不卡的大黄色大毛片视频在线观看| 亚洲最大成人手机在线| 日韩成人伦理影院| 美女主播在线视频| 久久午夜福利片| 另类亚洲欧美激情| 亚洲精品日韩在线中文字幕| 成人漫画全彩无遮挡| 久久精品国产亚洲av天美| 制服丝袜香蕉在线| 亚洲av一区综合| 99久久人妻综合| 国产精品福利在线免费观看| av卡一久久| 免费av观看视频| a级毛片免费高清观看在线播放| 国产欧美日韩精品一区二区| 亚洲内射少妇av| 亚洲三级黄色毛片| 国产亚洲91精品色在线| 男人添女人高潮全过程视频| 日本猛色少妇xxxxx猛交久久| 国产一区有黄有色的免费视频| 亚洲综合色惰| 夜夜爽夜夜爽视频| 高清av免费在线| 免费少妇av软件| 国产一区亚洲一区在线观看| 午夜免费观看性视频| 中文精品一卡2卡3卡4更新| 在线观看人妻少妇| 国产男人的电影天堂91| 欧美97在线视频| 久久久久久久久大av| 亚洲不卡免费看| 国产成人91sexporn| av线在线观看网站| 国产成人91sexporn| 精品人妻一区二区三区麻豆| 国产男人的电影天堂91| 青春草视频在线免费观看| 久久精品久久久久久噜噜老黄| 中文字幕亚洲精品专区| 国产精品av视频在线免费观看| 国产欧美日韩一区二区三区在线 | 极品教师在线视频| 少妇丰满av| 亚洲av在线观看美女高潮| 国内揄拍国产精品人妻在线| 别揉我奶头 嗯啊视频| 精品久久久久久电影网| 亚洲天堂国产精品一区在线| 精品午夜福利在线看| 99热这里只有是精品50| 激情五月婷婷亚洲| 国产精品不卡视频一区二区| 免费不卡的大黄色大毛片视频在线观看| 久久久久久久大尺度免费视频| 91久久精品电影网| 国产成人免费无遮挡视频| 久久久精品免费免费高清| 少妇人妻 视频| 熟女人妻精品中文字幕| 五月开心婷婷网| 精品午夜福利在线看| 美女视频免费永久观看网站| 成人毛片60女人毛片免费| 久久久久久久精品精品| 精品久久久噜噜| 少妇人妻久久综合中文| 久热久热在线精品观看| 欧美bdsm另类| 中文在线观看免费www的网站| av又黄又爽大尺度在线免费看| 成人美女网站在线观看视频| 精品久久久久久久久av| 国产成人一区二区在线| 免费观看无遮挡的男女| 夫妻性生交免费视频一级片| 亚洲四区av| 纵有疾风起免费观看全集完整版| 精品人妻熟女av久视频| 欧美亚洲 丝袜 人妻 在线| 免费看日本二区| 精品久久久久久久久亚洲| 成人亚洲精品av一区二区| 日本wwww免费看| 国产成人精品一,二区| 国产成人精品婷婷| 亚洲精华国产精华液的使用体验| 黄色怎么调成土黄色| 大香蕉久久网| 亚洲婷婷狠狠爱综合网| 三级国产精品欧美在线观看| 三级国产精品片| 日本与韩国留学比较| 久久99热6这里只有精品| 亚洲成人中文字幕在线播放| 国产综合懂色| 高清av免费在线| 秋霞伦理黄片| 国产成人精品婷婷| 在线观看av片永久免费下载| 午夜免费鲁丝| 成人免费观看视频高清| 国产免费视频播放在线视频| 岛国毛片在线播放| 日日摸夜夜添夜夜爱| 久久女婷五月综合色啪小说 | 高清毛片免费看| 久久99蜜桃精品久久| eeuss影院久久| 好男人视频免费观看在线| 伊人久久精品亚洲午夜| 亚洲精品国产色婷婷电影| 国产探花极品一区二区| 日韩三级伦理在线观看| 99热这里只有精品一区| 国产高清不卡午夜福利| 在线观看美女被高潮喷水网站| av在线老鸭窝| 在线观看一区二区三区| 精品人妻熟女av久视频| h日本视频在线播放| 黑人高潮一二区| 亚洲国产成人一精品久久久| 久久久亚洲精品成人影院| 美女内射精品一级片tv| 久久鲁丝午夜福利片| 国产黄片视频在线免费观看| 午夜免费男女啪啪视频观看| 国产一区亚洲一区在线观看| 熟妇人妻不卡中文字幕| 久久99热这里只有精品18| 免费黄频网站在线观看国产| 18禁裸乳无遮挡动漫免费视频 | 久久久国产一区二区| 18+在线观看网站| 日本爱情动作片www.在线观看| 亚洲成人av在线免费| 国产老妇伦熟女老妇高清| 午夜日本视频在线| 又爽又黄无遮挡网站| 日韩精品有码人妻一区| 成人欧美大片| 91在线精品国自产拍蜜月| 久久女婷五月综合色啪小说 | 亚洲经典国产精华液单| 色婷婷久久久亚洲欧美| 蜜臀久久99精品久久宅男| av线在线观看网站| 精品久久国产蜜桃| 91狼人影院| 好男人视频免费观看在线| 国内少妇人妻偷人精品xxx网站| 天堂俺去俺来也www色官网| 草草在线视频免费看| 夜夜爽夜夜爽视频| 一级二级三级毛片免费看| 色综合色国产| 亚洲成人久久爱视频| av国产免费在线观看| 国产一区有黄有色的免费视频| 99九九线精品视频在线观看视频| h日本视频在线播放| 亚洲国产色片| 久久精品久久久久久噜噜老黄| 观看免费一级毛片| 中国国产av一级| 美女脱内裤让男人舔精品视频| 丝袜脚勾引网站| 亚洲精品成人久久久久久| 午夜精品国产一区二区电影 | 校园人妻丝袜中文字幕| 免费av观看视频| 国产精品精品国产色婷婷| 午夜视频国产福利| 亚洲国产欧美在线一区| 精品亚洲乱码少妇综合久久| 哪个播放器可以免费观看大片| 三级国产精品片| 久久久午夜欧美精品| 97精品久久久久久久久久精品| 男的添女的下面高潮视频| 男人爽女人下面视频在线观看| 三级国产精品欧美在线观看| 日本欧美国产在线视频| 亚洲欧美一区二区三区黑人 | 99久久精品国产国产毛片| 亚洲真实伦在线观看| 伊人久久国产一区二区| 久久久精品欧美日韩精品| 人人妻人人澡人人爽人人夜夜| 狂野欧美激情性bbbbbb| 亚洲av成人精品一二三区| 美女被艹到高潮喷水动态| 国产成人a∨麻豆精品| 国产爱豆传媒在线观看| 亚洲图色成人| 成人特级av手机在线观看| 国产高清不卡午夜福利| 性色av一级| 亚洲一级一片aⅴ在线观看| 久久久久久久久久成人| 亚洲四区av| 黄片wwwwww| 免费人成在线观看视频色| 色播亚洲综合网| 欧美精品一区二区大全| 别揉我奶头 嗯啊视频| 成人特级av手机在线观看| 成人亚洲欧美一区二区av| 97精品久久久久久久久久精品| 久久久久久伊人网av| 中国美白少妇内射xxxbb| 亚洲最大成人手机在线| 免费观看a级毛片全部| 亚洲丝袜综合中文字幕| 午夜福利视频精品| 亚洲精品乱码久久久久久按摩| 熟妇人妻不卡中文字幕| 嫩草影院精品99| 美女脱内裤让男人舔精品视频| 女人被狂操c到高潮| 在线观看三级黄色| 精品酒店卫生间| 狂野欧美白嫩少妇大欣赏| 亚洲av.av天堂| 国产伦在线观看视频一区| 五月天丁香电影| 精品国产乱码久久久久久小说| 免费人成在线观看视频色| 91精品伊人久久大香线蕉| 免费高清在线观看视频在线观看| 国产色爽女视频免费观看| 久久6这里有精品| 在线观看av片永久免费下载| 新久久久久国产一级毛片| 超碰av人人做人人爽久久| 亚洲精品aⅴ在线观看| 亚洲精品一区蜜桃| 18禁在线播放成人免费| 1000部很黄的大片| 免费黄网站久久成人精品| av在线天堂中文字幕| 国产免费福利视频在线观看| 亚洲美女搞黄在线观看| 舔av片在线| 麻豆久久精品国产亚洲av| 成年av动漫网址| 18禁在线播放成人免费| 国内精品美女久久久久久| 亚洲不卡免费看| 亚洲av电影在线观看一区二区三区 | 一级毛片 在线播放| 久久久精品免费免费高清| 青青草视频在线视频观看| 最近中文字幕2019免费版| 久久久久网色| 一区二区三区精品91| 日日啪夜夜爽| 2018国产大陆天天弄谢|