• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    RNAGCN:RNA tertiary structure assessment with a graph convolutional network

    2022-11-21 09:32:26ChengweiDeng鄧成偉YunxinTang唐蘊芯JianZhang張建
    Chinese Physics B 2022年11期
    關(guān)鍵詞:王駿張建

    Chengwei Deng(鄧成偉) Yunxin Tang(唐蘊芯) Jian Zhang(張建)

    Wenfei Li(李文飛)1,2, Jun Wang(王駿)1,2, and Wei Wang(王煒)1,2,?

    1Collaborative Innovation Center of Advanced Microstructures,School of Physics,Nanjing University,Nanjing 210008,China

    2Institute for Brain Sciences,Nanjing University,Nanjing 210008,China

    RNAs play crucial and versatile roles in cellular biochemical reactions.Since experimental approaches of determining their three-dimensional (3D) structures are costly and less efficient, it is greatly advantageous to develop computational methods to predict RNA 3D structures. For these methods, designing a model or scoring function for structure quality assessment is an essential step but this step poses challenges. In this study, we designed and trained a deep learning model to tackle this problem. The model was based on a graph convolutional network(GCN)and named RNAGCN.The model provided a natural way of representing RNA structures, avoided complex algorithms to preserve atomic rotational equivalence,and was capable of extracting features automatically out of structural patterns. Testing results on two datasets convincingly demonstrated that RNAGCN performs similarly to or better than four leading scoring functions.Our approach provides an alternative way of RNA tertiary structure assessment and may facilitate RNA structure predictions. RNAGCN can be downloaded from https://gitee.com/dcw-RNAGCN/rnagcn.

    Keywords: RNA structure predictions, scoring function, graph convolutional network, deep learning, RNApuzzles

    1. Introduction

    RNAs play crucial and versatile roles in cellular biochemical reactions, such as encoding, decoding, catalysis,[1]gene regulations,[2]and others. These functions are closely related to the three-dimensional (3D) structures of the RNAs.To determine RNA 3D structure, experimental technics, including cryo-electron microscopy, x-ray crystallography, and nuclear magnetic resonance (NMR) spectroscopy are usually employed. Since these technics are costly and inefficient,computational methods have been developed to predict RNA structures.[3–19]Computational algorithms usually include two steps: (i) generating structural candidates and (ii) selecting structures most likely to be the native. The second step requires a good scoring function that can assess the quality of the candidates.

    Traditional scoring functions[20,21]are physics-based or knowledge-based, to name a few, Rosetta,[4,22]3dRNAscore,[23]RASP,[24]RNA KB potential,[25]DFIRERNA,[26]and rsRNASP.[27]Among them, 3dRNAscore is an all-atom statistical potential that combines distance-dependent and dihedral-dependent energies;[23]it is more efficient than other potentials in recognizing and ranking native state from a pool of near-native decoys. The rsRNASP potential is composed of short- and long-ranged energies, distinguished by residue separation along sequence;[27]extensive tests showed that it has higher or comparable performance against other leading potentials, dependent on specific testing datasets. In recent years, machine learning approaches achieved great success in many fields, including computer vision, natural language modelling,[28,29]medical diagnosis,[30]physics,chemistry, computational biology, and so on.[31–34]Inspired by these successes, our group developed a scoring function for the assessment of RNA tertiary structure based on a three-dimensional convolutional network and named it RNA3DCNN.[35]

    Applications of graph convolutional network(GCN)[36–41]to represent molecular structures have been quite successful.[42–47]For example, Foutet al.[43]modelled proteins as graphs at residue level and then predicted protein interfaces. Federicoet al.[44]introduced GCNs (GraphQA)to tackle the problem of quality assessment of protein structures, and showed that using only a few features will result in state-of-the-art results. Soumyaet al.[45]developed ProteinGCN for protein structures assessment; they built graphs to model spatial and chemical relations between pairs of atom.Zheet al.[46]built a GCN network, named GraphCPI, which aggregated chemical context of protein sequences and structural information of compounds to tackle the problem of compound–protein interaction. Huanget al.[47]introduced a model called GCLMI based on GCN and an auto-encoder to predict IncRNA–miRNA interactions.

    Inspired by the notion that a graph provides a more natural representation of 3D RNA structures and hence may bring better assessing performance, in this study, we upgrade our previous model by changing the convolutional network to a GCN network. Specifically, we developed a network model based on GCN, named RNA graph convolutional network(RNAGCN),to perform quality assessment of RNA structures.We trained and tested the model and compared the results with several leading scoring functions.

    2. Materials and methods

    In this section,we first introduce the input and output of RNAGCN, and then the architecture of the model. Next, we present the datasets used for training and evaluation,followed by a description of the metrics used to evaluate the performance of the model. At last,we describe the loss function and the training procedures.

    2.1. Graph representation of RNA structures

    RNA structures can be naturally represented as graphs,with nodes modeling atoms and edges modeling their relative spatial position in three-dimensional space. Since RNA sizes vary, we split an input RNA into many “l(fā)ocal environments”and converted each into a graph. Specifically, for theithnucleotide along the sequence, we defined it as the central nucleotide of its “l(fā)ocal environment” constructed by including all its nucleotide neighbors. A nucleotide was defined to be a neighbor if any of its heavy atoms (other than a hydrogen atom) was within a threshold spatial distance of any heavy atoms of the central nucleotide. The distance threshold was set to be 14 ?A.This is to construct a large enough local environment for the central nucleotide.With this design,the model can be scaled to handle RNAs of arbitrary size. Thus, for an RNA of lengthNs,we constructedNslocal environments represented byNsgraphs.

    In general,a graph is defined as a set

    whereVdenotes the set of nodes andEdenotes the set of edges,respectively.X ∈R|V|×NandU ∈R|?|×Eare the feature matrices ofVandE,respectively.N,Eare the number of nodes and number of edges,respectively.

    All heavy (non-hydrogen) atoms in a local environment were treated as nodes. A node was represented as a one-hot vector of lengthNt,whereNtis the total number of atom types,which is 54,based on AMBER99SB force field.

    Edges were defined between neighboring atoms. Each atom was connected by an edge with itsKnearest atoms in space, whereKwas set to be 14. Edges were directed and each included five features. The first was the spatial distance betweenSandT, which was one of its neighbors. The 2nd,3rd,and 4th were direction features,which were the three projections of the unit vector pointing fromStoTin the internal coordinate system centered onS,i.e.,

    whereAdenotes the ‘C1’ atom andBthe ‘C5’ atom in a nucleotide. The fifth feature had a value of 1 or 0,depending on whether there is a chemical bond betweenSandT.

    2.2. Output of RNAGCN:quality score

    The graphs representing the local nucleotide environment constructed for each nucleotide from its neighbors were used as input to the GCN model. The model output was a scalar score indicating the quality of the input. During training,the output score measured the difference of local environment from the ground truth, i.e., the RMSD of the input structure with respect to the experimental one. During the inferring operation,the network output was the predicted score indicating the RMSD of the input versus the experimental structure that in this operation was an unknown to the network. The scores thus obtained for theNsinput environments of an RNA were averaged to get the final score,which reflected the overall difference of the input structure to the experimental one, with a lower score corresponding to a higher input quality (a better approximation of the unknown experimental structure).

    2.3. Architecture of RNAGCN

    RNAGCN is a deep learning model based on GCN that is used to extract features from 3D structures of RNA. The model architecture is shown in Fig. 1. The input to the model is an RNA local graph,converted from an RNA local environment. Below the input layer,there are five serially connected graph convolution layers that operate on the graph sequentially.Residual modules and skip connections were adopted to solve the gradient vanishing problem often seen in deep neural networks.[38,40]Serially connected below the GCN layers,there is a convolution layer with a 1×1 kernel, followed by a graph global max pooling layer, and then the final layer of a fully-connected network that outputs a score as prediction.The model contains about 378k parameters in total.

    In the five GCN layers, graph convolution operations were used to update the nodes’representations iteratively. The details of the operations are shown on the right side of Fig.1.

    Fig.1. Overview of the RNAGCN model. The input is an RNA local graph. The network contains five graph convolution layers,a convolution layer with a 1×1 kernel,a graph global max pooling layer,and a fully-connected network that outputs a score as prediction. The right panel presents details of feature operations using graph convolution with residual connections.

    Following the GCN layers, a convolution layer with a 1×1 kernel was used to perform the aggregation of information fromLGCN nodes with different weights. These weights were learned from data during training. As a result,Xwas reduced fromRL×|V|×NtoR|V|×N.

    A global max pool layer was used to mix nodes representations. Max values ofX ∈R|V|×Nalong|V|were computed one by one. As a result,Xwas further reduced,fromR|V|×NtoRN.

    Finally,a fully-connected network was used to reduceXfromRNtoR1.It included a layer of size 64 with ReLU activation function,and a layer of size 1 without activation function.

    Notably, the representation of edges was kept static during graph convolution,according to the generally accepted notion that updating nodes’features alone is usually sufficient to achieve reasonable model performance.

    2.4. Datasets

    We built our datasets based on the non-redundant set of RNA 3D Hub[48](http://rna.bgsu.edu/rna3dhub/nrlist,Release 3.102, 2019-11-27). RNAs forming complex with proteins or other molecules were removed; only pure RNA structures were kept. Short RNAs with length smaller than eight nucleotides were also removed. The remaining 610 RNAs were split into training, validation and testing datasets. The infernal program[49]was used to ensure that no RNA in the testing dataset belonged to the same RFAM family as in the training and validation datasets. The resulting three datasets contained 426, 92, and 92 RNAs, respectively. Hereafter, the testing dataset of 92 RNAs is referred to as Test-I.

    RNA decoys are needed to train and test scoring functions. From the experimental structure of each RNA, we generated the corresponding decoys with high-temperature molecular dynamics (MD) simulations, carried out using Gromacs.[50]In brief, each RNA molecule was solved in a TIP3P water box,and metal ions were added to neutralize the system. The system was then subjected sequentially to energy minimization, NVT equilibration, and NPT equilibration at 300 K.Finally,high temperature MD was used to denature the structure by gradually increasing the temperature from 300 K to 660 K.The force field was AMBER99SB.Each simulation lasted for 10 ns. For each trajectory,we calculated the RMSD of each frame with respect to the corresponding experimental structure and randomly selected 200 decoys in the RMSD range from 0 to 20 ?A. The decoys, together with all the experimental structures,constructed the training,validation,and testing datasets,as summarized in Table 1.

    Fig.2. Distributions of(a)RMSD and(b)length for four datasets. Note that RNA 6qkl in the training set is not shown,for its length is too long(1158 nucleotides).

    We prepared another testing dataset based on the RNApuzzles-standardized dataset,[51]which included 22 RNAs.Again, for each RNA in this set, the same denaturing procedure was carried out by high-temperature MD to generate decoys, and 200 decoys were randomly selected. The resulting testing dataset,referred to as Test-II hereafter,contained 4422 RNA structures.

    The distribution of RMSDs and lengths for the abovementioned datasets are given in Fig.2.

    Table 1. Information of datasets.

    2.5. Evaluation metrics

    Enrichment score (ES) and Pearson correlation coefficient(PCC)were used to evaluate the performance of scoring function.

    Enrichment score(ES)[23,25–27,35,52]was defined as

    whereStop10%andRtop10%indicated the best-scored 10% decoys and the 10% with the lowest RMSDs, respectively.Ndecoyswas the number of decoys associated with the RNA of concern. ES values ranged from 0 to 10, with larger value indicating better performance.

    The Pearson correlation coefficient (PCC)[24,27,52]indicates the magnitude of linear correlation between two variables and is defined as

    whereSiandRiare the scores given by the scoring function and the RMSD of the structure, respectively. PCC assumes a value between 0 and 1.

    2.6. Loss function

    The loss function, used as the optimizing target of the model,was defined as

    whereRiandSiwere RMSD and the predicted score of theithlocal environment,respectively.Nenvswas the total number of local environments in the whole training dataset. In this study,it amounts to about 107.

    2.7. Training strategies

    All the RNAs were partitioned into local environments,which were then randomly shuffled and fed to the network in batches. The batch size per GPU was set to 64 and the total batch size was 384, since 6 GPUs were used in parallel computation. Python modules Deep Graph Library and Pytorch.Distributed were used for model construction and computation parallelization,respectively. Mini-batch gradient descent algorithm was employed to train the model.Specifically,the Adam optimizer with parametersβ1=0.9,β2=0.98 was used. The initial learning rate was 0.004 and it would be decreased by half if the training loss no longer dropped for 4 epochs.

    3. Results

    In this section,we present the results of our model experiments and compare the performance of our model with four statistical potentials, including RASP, Rosetta, 3dRNAscore,and rsRNASP, which are currently the most popular in the field.

    3.1. Performance on Test-I

    Table 2 shows the performance of the five models in assessing the quality of structures in the Test-I dataset, which contained 92 RNAs and 200 decoys associated with each.We used two criteria, labeled as Top-1 and Top-5, to reflect whether the experimental RNA was ranked the first,or ranked among the best five, respectively. As judged by Top-1, 3dRNAscore and rsRNASP performed best;they identified 91 and 86 out of 92 RNAs,respectively. RNAGCN(our model)performed as a third (79/92). Judged by the criterion Top-5, all models performed similarly. According to comparisons based on these two criteria, our model is the third best among all.This outcome is tentatively attributed to the strict criteria we used to construct the testing set: we made sure there are no overlapping of RFAM families between the testing set and the training datasets.

    Table 2. Performance of five models on Test-I.

    Fig.3.(a)Experimental structure of RNA 5dcv.(b)Score-RMSD plots,showing the correlation between the scores predicted by the model and the RMSDs of the structures. The purple crosses mark the experimental structures, orange crosses indicate the structures that were scored better than the experimental ones, and red crosses mark the best-scored structures if they were not the experimental one.

    Table 2 also presents the average values for ES and PCC that indicate the strength of correlations between the ground truth and prediction.Our model is superior to the others by the ES measure, close second best behind RASP by the average PCC, and clearly superior by the measure sensitive to nearnative structures (PCC with RMSD<4 ?A). The latter result indicates that our model excels at discriminating small structural changes with respected to the native one.

    In Fig. 3(b) we present the score-RMSD plots for each of the five models, computed using RNA-5dcvas an example(Fig.3(a)),which is a 95-nucleotid-long fragment fromP.horikoshiiRNase.[53]The data show that RASP and Rosetta failed to identify the experimental structure as the best one.In contrast,3dRNAscore,rsRNASP and our model ranked the experimental one as the best structure(as indicated by overlapping symbols for best score and native). Moreover,our model appeared to show a better score-RMSD correlation, particularly at small RMSD ranges.For a further proof,we calculated the average PCC values for RNAs within the 0 to 4 ?A range of RMSD and found them to be 0.77,0.47,-0.65,0.61,and 0.70 for RNAGCN, RASP, Rosetta, 3dRNAscore, and rsRNASP,respectively. By this measure, our model clearly exhibited a significantly higher correlation between scores and RMSDs.This result is also consistent with that bottom line in Table 2,which shows that our model gives the best correlation in the small RMSD range. Moreover, for this specific RNA,the ES values of RNAGCN, RASP, Rosetta, 3dRNAscore, and rsRNASP are 5.0,3.0,1.0,0.5,and 5.0,respectively,also consistent with the ES result presented in Table 1.

    The ES, PCC, and score-RMSD plots for each RNA in Test-I are listed in the supplementary material.

    3.2. Performance on Test-II

    We also analyzed the performance of the five models on the Test-II testing dataset, which is composed of 22 RNAs taken from RNA-puzzles-standardized datasets. The result is summarized in Table 3.

    Table 3 shows that 3dRNAscore identified all 22 experimental structures correctly and our model identified just one less, 21 out of 22 RNAs, a close second. Judged by the ES values, our model is 6.5, which is very closed to the best 6.7(rsRNASP).The average PCC of our model is 0.88,also very close second to the best value 0.90 (rsRNASP). Within the small RMSD range (<4 ?A), the PCC value of our model is again close to the best 0.76 (rsRNASP). Detailed results for each RNA in Test-II can be found in the supplementary material.

    Table 3. Performance of five models on Test-II.

    Fig. 4. (a) Experimental structure of puzzle-2 in the RNA-puzzles dataset. (b) Score-RMSD plots, showing the correlation between the predicted scores and the RMSDs of the structure. The purple, orange,and red crosses are defined in the same way as in Fig.3.

    Figure 4(b)shows the score-RMSD plots for puzzle-2 in Test-II.The pattern of performances is similar to that of Test-I in Fig.3. Three models, including 3dRNAscore, rsRNASP,and ours,correctly identified the experimental structure out of decoys. For this specific RNA, the ES value of our model is 9.5,very closed to perfect value 10;and the PCC value is 0.95.Our model and RASP appear to show a good correlation between scores and RMSDs, while Rosetta, 3dRNAscore, and rsRNASP feature a big gap in scores between the experimental structure and the decoys. The score-RMSD plots for other RNAs in Test-II are generally similar to Fig.4 and listed in full in the supplementary material.

    We compiled a detailed ranking of the five models for Test-II and presented the results as a colormap in Fig.5. For each RNA in Test-II and each scoring function, we selected the best-scoredNb(a pre-set variable, “b” indicates best) decoys and computed their average RMSD. We then ranked five scoring functions based on these RMSDs and represented the ranks with different colors, with lighter colors indicating higher ranks. It can be seen that RNAGCN has the largest area of light colors(light yellow and gold),indicating that our model gives the lowest RMSDs more often than the others.

    Following the usual practice in the field of machine learning,we plotted receiver operating characteristic(ROC)curves on difference choices ofNb, used as a variable threshold to select structures as the best predicted ones. For example, ifNbwas 10, we trusted the best scored 10 structures ranked by the model as the predicted native structures. To test the model performance under different choices ofNb,we first defined a native ensemble as all structures with RMSD smaller than 2 ?A,and then defined all structures within the native ensemble as positive samples and the others as negative. These structures and the associated binary classification labels were used to construct the ground-truth dataset, following conventions in machine learning field.During predictions,the bestNbstructures given by model were treated as the predicted positives,while the leftover was treated as the predicted negatives.Under varying choices ofNb, we calculated the true-positive rate (TPR) and false-positive rate (FPR) and plotted their relationship as ROC curves in Fig. 6. It can be seen that the ROC curve of our model almost envelopes (i.e., maximizes)the other four curves. Quantitatively, the areas under ROC(AUC) of our model for Test-I and Test-II are 0.97 and 0.89,respectively; both represent the best performance among the all scoring functions.

    Fig. 5. Ranking of five models based on the average RMSDs on the best-scored Nb structures, where Nb is a variable plotted on the x-axis.Each row corresponds to one RNA in Test-II(a total of 22,note their labels are not continuous in the standard dataset). The ranks are indicated by the color bar on the right;the lighter the color,the higher the rank.

    Fig.6. ROC curves of five scoring functions on different choices of threshold Nb. Panels (a) and (b) are results obtained for the datasets Test-I and Test-II,respectively.

    4. Discussion and conclusions

    Applications of deep learning technology in molecular representation learning have achieved impressive progress in recent years. Inspired by previous works,we explored the application of graph convolutional network in RNA 3D structure assessment tasks. We compiled two datasets with MD simulations and trained a GCN neural network. We tested the model and evaluated its performance using four leading scoring functions. The testing dataset Test-I contained 92 RNAs. Dataset Test-II contained 22 RNAs,a subset of the RNA-puzzles standardized dataset.

    For both testing datasets,the ability of our GCN model to identify experimental structures closely approached the best–3dRNAscore on Test-I and rsRNASP on Test-II,respectively.The ES metric,measuring the overlap of the near-native structures and the best scored structures, indicated that our model is the best one for distinguish near-native structures (<4 ?A)on Test-I while the second one on Test-II. Additionally, we ranked five models based on average RMSD of the best scored structures, and compared their ROC curves and AUCs; both experiments ranked our model as superior to the others.

    It is interesting to compare our machine learning model with the statistical potentials based on inverse-Boltzmann equation,particularly,3dRNAscore and rsRNASP.According to the presented results, our model seems to perform slightly worse in identifying native structures (Tables 2 and 3) and slightly better in comparison of AUCs. This outcome may be tentatively attributed to the strict criteria we used to construct the testing set: we made sure there are no overlapping of RFAM families between the testing set and the training datasets. Besides, we did not train the model over and over to increase the metrics,since too much training might lead to over-fitting and then decrease generalization ability, even the training and testing datasets were strictly separated.

    In general, the advantages of statistical potentials based on inverse-Boltzmann equation are their clear physical picture,good generalization ability to unseen structures,and fast computation speed. In contrast,machine learning based potentials are black-boxes and have unclear generalization ability and arguably slow speed.However,the advantages of machine learning approaches are also prominent. First,it needs not years of studies looking for proper energy forms and choice of reference states.[52]For RNA structures,it has taken people a dozen years to achieve the current performance. While a similar performance can be achieved by simply taking a general deep network(with slightly modifying the input and output layers)and training it properly. This advantage becomes more prominent if people need to migrate model to new molecules. Second,deep neural networks have found physical patterns similar to those found by human,and have the potential to find new patterns not-known yet. Machine learning based scoring models deserve to be studied.

    A recently introduced geometric deep learning approach named ARES[54]showed good performance at scoring RNA structures, but unfortunately, it is not yet possible to directly compare it to ours. The authors of ARES have not released their trained model, and we failed to reproduce their results by attempting to rebuild the model from downloaded codes and training the same network with their datasets. Moreover,ARES was trained on datasets generated with FARFAR2,different from the datasets we generated from MD simulations.A fair comparison of these two models requires identical data or data sampled from independently identical distribution for training and testing.

    Our model has several advantages. First, it is natural to represent the geometric and topologic information about RNA tertiary structures as a graph. Second, because a graph automatically provides geometric invariant representation regarding atomic translation and rotation, there is no need to design complex convolution operations such as those used in ARES.Third,because our model can directly learn from spatial patterns of atoms on its own, it requires, unlike physicsbased scoring functions,no prior physical knowledge. Therefore, our approach can be easily extended to other molecular systems with little modification. Forth, the design of splitting structures into local environments makes the model scalable,enabling us to treat RNAs of arbitrary size. At last,this study showed that our model performed well particularly for near-native structures. We tentatively attributed this feature to the graph representation of tertiary structures, for both graph topology and edge features are sensitive to the changes in positions of atoms.

    There are two notable limitations of the present work.One limit is computational, stemming from the huge memory consumption of graphs and the massive computations involved in graph convolution operations. As both factors limit the size of the graph,and the number of neighbor atoms around the central nucleotide,they may raise problems handling very long-range interactions that are relevant. The second is algorithmic,as in the current version of the model only node features in the network were updated while edge features were kept static. Presumably,updating both in an upgraded version may slightly improve model performance.

    Acknowledgements

    This study was funded by the National Natural Science Foundation of China(Grant Nos.11774158 to JZ,11934008 to WW,and 11974173 to WFL).The authors acknowledge High Performance Computing Center of Advanced Microstructures,Nanjing University for the computational support.

    猜你喜歡
    王駿張建
    護苗
    中國火炬(2024年4期)2024-04-17 11:27:44
    趕海記憶
    鷸蚌相爭,漁人得利
    Diffusion of nucleotide excision repair protein XPA along DNA by coarse-grained molecular simulations?
    豆腐里的愛
    本期焦點人物:張建一
    綠色中國(2019年13期)2019-11-26 07:10:50
    《CT成像:基本原理、偽影與誤區(qū)》已出版
    書寫精氣神
    榮譽雜志(2015年2期)2015-07-08 14:23:56
    肩關(guān)節(jié)生物力學(xué)
    真情假意
    故事林(2007年12期)2007-05-14 15:37:51
    男女下面插进去视频免费观看| 亚洲四区av| 伦理电影大哥的女人| 欧美亚洲日本最大视频资源| 高清不卡的av网站| 欧美精品国产亚洲| 大香蕉久久成人网| 国产免费现黄频在线看| 超碰97精品在线观看| 国产人伦9x9x在线观看 | 中文精品一卡2卡3卡4更新| 精品国产露脸久久av麻豆| 亚洲国产看品久久| 老司机影院成人| 在线观看国产h片| 亚洲美女视频黄频| 亚洲,欧美精品.| 妹子高潮喷水视频| 99久久综合免费| 一级毛片 在线播放| 亚洲欧美一区二区三区黑人 | 午夜免费观看性视频| 最近的中文字幕免费完整| 亚洲av电影在线观看一区二区三区| 欧美bdsm另类| 欧美日韩综合久久久久久| 另类亚洲欧美激情| 丁香六月天网| 水蜜桃什么品种好| av在线播放精品| 免费av中文字幕在线| 国产一区二区在线观看av| 国产欧美亚洲国产| av在线老鸭窝| 麻豆精品久久久久久蜜桃| 久久午夜福利片| av免费在线看不卡| 亚洲伊人久久精品综合| av福利片在线| 成人毛片a级毛片在线播放| 日韩在线高清观看一区二区三区| 国产在线一区二区三区精| a级毛片黄视频| 亚洲国产精品一区三区| 欧美国产精品一级二级三级| 美女主播在线视频| 免费观看在线日韩| 男女无遮挡免费网站观看| 99热网站在线观看| 久久久国产精品麻豆| 免费在线观看完整版高清| 成人午夜精彩视频在线观看| 精品国产一区二区三区四区第35| 在线亚洲精品国产二区图片欧美| 成年人午夜在线观看视频| av电影中文网址| 精品少妇内射三级| 久久久久人妻精品一区果冻| 国产一区二区在线观看av| 成人毛片a级毛片在线播放| 久久久久精品人妻al黑| 午夜福利视频在线观看免费| 水蜜桃什么品种好| 男男h啪啪无遮挡| 成人漫画全彩无遮挡| 十分钟在线观看高清视频www| 国产极品天堂在线| 免费高清在线观看日韩| 精品一品国产午夜福利视频| 女人高潮潮喷娇喘18禁视频| 另类精品久久| 国产精品久久久久成人av| av一本久久久久| 波多野结衣一区麻豆| 欧美精品人与动牲交sv欧美| 黄色配什么色好看| 免费少妇av软件| 9热在线视频观看99| 一二三四中文在线观看免费高清| 国产激情久久老熟女| 啦啦啦在线观看免费高清www| 91午夜精品亚洲一区二区三区| 国产精品一区二区在线观看99| 国产午夜精品一二区理论片| www.自偷自拍.com| 母亲3免费完整高清在线观看 | 999久久久国产精品视频| 婷婷色综合www| 久久99热这里只频精品6学生| 免费av中文字幕在线| 菩萨蛮人人尽说江南好唐韦庄| freevideosex欧美| 伊人亚洲综合成人网| 少妇被粗大的猛进出69影院| 久久久久久久国产电影| 激情视频va一区二区三区| 2021少妇久久久久久久久久久| 亚洲成人手机| 中文字幕最新亚洲高清| 国产成人精品婷婷| 国产97色在线日韩免费| 看非洲黑人一级黄片| 欧美精品人与动牲交sv欧美| 丝袜美腿诱惑在线| av有码第一页| 不卡视频在线观看欧美| 国产精品不卡视频一区二区| 久久久精品免费免费高清| 新久久久久国产一级毛片| 老汉色∧v一级毛片| 性色av一级| 999精品在线视频| 中文字幕精品免费在线观看视频| 午夜福利视频精品| 高清不卡的av网站| 国产av国产精品国产| 久久人人爽av亚洲精品天堂| 国产精品免费大片| av天堂久久9| 亚洲欧洲精品一区二区精品久久久 | 久久久国产欧美日韩av| 丝瓜视频免费看黄片| 久久久久久人妻| 亚洲精品中文字幕在线视频| 老汉色∧v一级毛片| 精品少妇一区二区三区视频日本电影 | 人妻人人澡人人爽人人| 亚洲国产精品一区三区| 亚洲国产精品一区三区| 久久人人爽av亚洲精品天堂| 丰满迷人的少妇在线观看| 国产精品欧美亚洲77777| av免费观看日本| 成人毛片a级毛片在线播放| 1024视频免费在线观看| 日本色播在线视频| 日韩av不卡免费在线播放| 青春草国产在线视频| 两个人免费观看高清视频| 日本黄色日本黄色录像| 精品一品国产午夜福利视频| 欧美人与性动交α欧美软件| 国产成人免费无遮挡视频| 9191精品国产免费久久| 亚洲国产最新在线播放| 美国免费a级毛片| 欧美国产精品va在线观看不卡| 亚洲一区中文字幕在线| 免费黄色在线免费观看| 一本久久精品| 在现免费观看毛片| 在线 av 中文字幕| 成年人午夜在线观看视频| 9热在线视频观看99| 国产免费一区二区三区四区乱码| 精品少妇内射三级| 大片免费播放器 马上看| 好男人视频免费观看在线| 9191精品国产免费久久| 久久韩国三级中文字幕| 中文字幕av电影在线播放| 黑人欧美特级aaaaaa片| 国产片内射在线| 亚洲欧洲日产国产| 亚洲内射少妇av| 免费av中文字幕在线| 搡老乐熟女国产| 国产高清国产精品国产三级| 夫妻午夜视频| 亚洲情色 制服丝袜| 青春草亚洲视频在线观看| 看免费av毛片| 在线观看免费高清a一片| 国精品久久久久久国模美| 亚洲一码二码三码区别大吗| 国产探花极品一区二区| videos熟女内射| 亚洲精品一区蜜桃| 欧美97在线视频| 亚洲久久久国产精品| 最近中文字幕2019免费版| 日韩av在线免费看完整版不卡| 2022亚洲国产成人精品| 国产成人精品久久二区二区91 | 国产成人精品无人区| 国产精品国产av在线观看| av视频免费观看在线观看| 久久久久精品人妻al黑| 国产精品不卡视频一区二区| 久久久久国产网址| 亚洲经典国产精华液单| 岛国毛片在线播放| 一级毛片黄色毛片免费观看视频| 纵有疾风起免费观看全集完整版| 少妇被粗大猛烈的视频| 日韩av免费高清视频| 欧美日韩国产mv在线观看视频| 韩国av在线不卡| 精品亚洲乱码少妇综合久久| 久久毛片免费看一区二区三区| 香蕉丝袜av| 国产精品无大码| 又粗又硬又长又爽又黄的视频| 丝袜美足系列| 人妻系列 视频| 午夜福利视频在线观看免费| 捣出白浆h1v1| 亚洲激情五月婷婷啪啪| 日韩中文字幕视频在线看片| 国产免费视频播放在线视频| 少妇 在线观看| 久久久久久人妻| 亚洲av在线观看美女高潮| 韩国高清视频一区二区三区| 自线自在国产av| 亚洲av综合色区一区| 亚洲av在线观看美女高潮| 丝袜喷水一区| 国产不卡av网站在线观看| 亚洲精品美女久久久久99蜜臀 | 99精国产麻豆久久婷婷| 免费观看在线日韩| 91国产中文字幕| 1024视频免费在线观看| 男人添女人高潮全过程视频| 在线观看一区二区三区激情| 亚洲第一青青草原| 中文字幕制服av| 97在线视频观看| 欧美日韩亚洲国产一区二区在线观看 | 美女高潮到喷水免费观看| 成人毛片a级毛片在线播放| 日韩av在线免费看完整版不卡| 日本色播在线视频| 国产精品人妻久久久影院| √禁漫天堂资源中文www| 啦啦啦视频在线资源免费观看| 一本久久精品| 少妇 在线观看| www.自偷自拍.com| 国产免费又黄又爽又色| 赤兔流量卡办理| 亚洲成人av在线免费| 久久久久久久久久人人人人人人| 满18在线观看网站| 亚洲色图 男人天堂 中文字幕| 最新的欧美精品一区二区| 久久精品久久久久久噜噜老黄| 亚洲国产欧美在线一区| 亚洲欧洲精品一区二区精品久久久 | 色网站视频免费| 制服诱惑二区| 视频区图区小说| 777米奇影视久久| 国产精品欧美亚洲77777| 高清av免费在线| 丰满迷人的少妇在线观看| 咕卡用的链子| 777久久人妻少妇嫩草av网站| 老汉色∧v一级毛片| 精品亚洲成国产av| 久久久欧美国产精品| 久久鲁丝午夜福利片| 久久精品国产亚洲av高清一级| 中文字幕另类日韩欧美亚洲嫩草| 2022亚洲国产成人精品| 国产成人91sexporn| 各种免费的搞黄视频| 制服人妻中文乱码| 99久久综合免费| www日本在线高清视频| 日韩中文字幕欧美一区二区 | 9色porny在线观看| 国产精品欧美亚洲77777| 欧美日韩视频高清一区二区三区二| 在线观看免费高清a一片| 亚洲欧美一区二区三区国产| 丝袜在线中文字幕| 91成人精品电影| 国产欧美亚洲国产| 在线看a的网站| 如日韩欧美国产精品一区二区三区| 成人手机av| 久久久久久伊人网av| 黄色怎么调成土黄色| av国产久精品久网站免费入址| 亚洲欧美日韩另类电影网站| 波多野结衣av一区二区av| 欧美激情 高清一区二区三区| 国产成人精品在线电影| 人成视频在线观看免费观看| 爱豆传媒免费全集在线观看| 街头女战士在线观看网站| 80岁老熟妇乱子伦牲交| 亚洲第一区二区三区不卡| 亚洲成国产人片在线观看| 午夜福利视频精品| 久久久久久久国产电影| 99国产精品免费福利视频| 欧美日韩一级在线毛片| 青青草视频在线视频观看| 老汉色∧v一级毛片| kizo精华| 亚洲精品日韩在线中文字幕| 欧美精品av麻豆av| 在线观看www视频免费| 亚洲,欧美精品.| 日产精品乱码卡一卡2卡三| 欧美国产精品一级二级三级| 水蜜桃什么品种好| 成人午夜精彩视频在线观看| 男的添女的下面高潮视频| 亚洲欧洲精品一区二区精品久久久 | 超碰97精品在线观看| 亚洲精品久久成人aⅴ小说| 免费女性裸体啪啪无遮挡网站| 最近最新中文字幕免费大全7| 性色avwww在线观看| tube8黄色片| 中文天堂在线官网| 国产精品免费大片| 亚洲欧美中文字幕日韩二区| 国产又爽黄色视频| 久久久久久久久久人人人人人人| 日日撸夜夜添| 国产精品久久久久成人av| 精品视频人人做人人爽| 免费高清在线观看视频在线观看| 高清不卡的av网站| 老女人水多毛片| 国产女主播在线喷水免费视频网站| 国产激情久久老熟女| 性色av一级| 黄色配什么色好看| 免费在线观看黄色视频的| 性色avwww在线观看| h视频一区二区三区| 午夜日本视频在线| 日本爱情动作片www.在线观看| 国产精品蜜桃在线观看| 国产高清不卡午夜福利| 精品午夜福利在线看| 免费黄色在线免费观看| 人妻一区二区av| 波多野结衣一区麻豆| 热re99久久国产66热| 丝袜喷水一区| 满18在线观看网站| 亚洲情色 制服丝袜| 狠狠精品人妻久久久久久综合| 成人亚洲欧美一区二区av| 亚洲成国产人片在线观看| 最近的中文字幕免费完整| 亚洲,欧美,日韩| 亚洲综合色惰| www日本在线高清视频| 另类亚洲欧美激情| 人人妻人人澡人人看| 久久国内精品自在自线图片| 精品国产一区二区三区四区第35| 这个男人来自地球电影免费观看 | 日韩三级伦理在线观看| 曰老女人黄片| a级片在线免费高清观看视频| 26uuu在线亚洲综合色| 国精品久久久久久国模美| 一级爰片在线观看| 亚洲精品日韩在线中文字幕| 我的亚洲天堂| 欧美成人午夜免费资源| 亚洲第一区二区三区不卡| 国产亚洲午夜精品一区二区久久| 国产亚洲一区二区精品| 精品人妻熟女毛片av久久网站| 亚洲av男天堂| 亚洲一区中文字幕在线| 中文字幕人妻丝袜一区二区 | 亚洲伊人久久精品综合| 中文乱码字字幕精品一区二区三区| 国产成人午夜福利电影在线观看| 欧美日韩精品成人综合77777| 水蜜桃什么品种好| 1024香蕉在线观看| 最黄视频免费看| av电影中文网址| 亚洲人成网站在线观看播放| 亚洲av国产av综合av卡| 色视频在线一区二区三区| 极品少妇高潮喷水抽搐| 精品人妻偷拍中文字幕| 毛片一级片免费看久久久久| 国产免费又黄又爽又色| 亚洲人成电影观看| 午夜福利视频精品| 免费少妇av软件| 日本午夜av视频| 美女国产高潮福利片在线看| 国产探花极品一区二区| 七月丁香在线播放| 在线观看www视频免费| 精品国产露脸久久av麻豆| 国产精品一区二区在线不卡| 久久精品国产a三级三级三级| a级毛片在线看网站| 久久国产亚洲av麻豆专区| 国产成人精品一,二区| 欧美精品一区二区免费开放| 啦啦啦在线免费观看视频4| 巨乳人妻的诱惑在线观看| 欧美日韩视频精品一区| 婷婷色综合www| 捣出白浆h1v1| 777米奇影视久久| 狠狠精品人妻久久久久久综合| 大话2 男鬼变身卡| 国产综合精华液| 丁香六月天网| 男女边摸边吃奶| av卡一久久| 午夜福利乱码中文字幕| 午夜av观看不卡| 亚洲av免费高清在线观看| 欧美日本中文国产一区发布| 爱豆传媒免费全集在线观看| 免费看av在线观看网站| 人人妻人人澡人人爽人人夜夜| 国产男女内射视频| 丝袜在线中文字幕| av网站免费在线观看视频| 精品少妇内射三级| 亚洲人成网站在线观看播放| 亚洲伊人久久精品综合| 水蜜桃什么品种好| 亚洲成人av在线免费| 久久久久国产网址| 久久精品亚洲av国产电影网| 大片免费播放器 马上看| 人妻 亚洲 视频| 搡女人真爽免费视频火全软件| 秋霞在线观看毛片| 中文字幕另类日韩欧美亚洲嫩草| 国产精品女同一区二区软件| 自线自在国产av| 1024视频免费在线观看| 日韩欧美精品免费久久| 免费在线观看完整版高清| 人人妻人人澡人人看| 麻豆乱淫一区二区| 最近最新中文字幕大全免费视频 | 日韩,欧美,国产一区二区三区| 日本猛色少妇xxxxx猛交久久| 国产精品无大码| 日本欧美国产在线视频| 国产精品二区激情视频| 少妇人妻久久综合中文| 激情视频va一区二区三区| 熟女电影av网| 国产av国产精品国产| 亚洲精品一二三| 日韩三级伦理在线观看| 欧美 日韩 精品 国产| 纵有疾风起免费观看全集完整版| 欧美日韩视频高清一区二区三区二| 久久久a久久爽久久v久久| 中文欧美无线码| 国产精品99久久99久久久不卡 | videosex国产| 中文字幕最新亚洲高清| 一个人免费看片子| 国产精品一国产av| 国产精品免费视频内射| 精品酒店卫生间| 91午夜精品亚洲一区二区三区| 黄片无遮挡物在线观看| 亚洲av免费高清在线观看| 午夜福利在线免费观看网站| xxxhd国产人妻xxx| 亚洲视频免费观看视频| 亚洲综合色惰| 亚洲国产成人一精品久久久| 国产精品国产三级专区第一集| 亚洲精品日本国产第一区| 欧美日韩一区二区视频在线观看视频在线| 黄色毛片三级朝国网站| 赤兔流量卡办理| 大码成人一级视频| 新久久久久国产一级毛片| 男人操女人黄网站| 久久久久国产精品人妻一区二区| 欧美激情高清一区二区三区 | 最近中文字幕高清免费大全6| 亚洲视频免费观看视频| 中国三级夫妇交换| 乱人伦中国视频| 成人手机av| 一边摸一边做爽爽视频免费| 交换朋友夫妻互换小说| 国产在视频线精品| 国产一区二区激情短视频 | 精品卡一卡二卡四卡免费| 国产精品香港三级国产av潘金莲 | 80岁老熟妇乱子伦牲交| 各种免费的搞黄视频| 香蕉精品网在线| 成人毛片a级毛片在线播放| 美国免费a级毛片| 黑丝袜美女国产一区| 啦啦啦在线免费观看视频4| 人人妻人人爽人人添夜夜欢视频| 日韩中文字幕视频在线看片| 亚洲图色成人| 99热国产这里只有精品6| 一级黄片播放器| 精品人妻偷拍中文字幕| 国产精品免费视频内射| 欧美精品人与动牲交sv欧美| 曰老女人黄片| 男的添女的下面高潮视频| 亚洲精品日本国产第一区| 精品国产露脸久久av麻豆| 国产在视频线精品| 亚洲人成77777在线视频| 国产精品久久久久久av不卡| 女人久久www免费人成看片| 亚洲色图综合在线观看| 91精品三级在线观看| 在线观看美女被高潮喷水网站| 久久久久精品人妻al黑| 人人妻人人添人人爽欧美一区卜| 1024香蕉在线观看| 成人影院久久| 国产精品国产av在线观看| 国产激情久久老熟女| 欧美97在线视频| 国产男人的电影天堂91| 亚洲精品久久成人aⅴ小说| 大片电影免费在线观看免费| 亚洲第一av免费看| 免费不卡的大黄色大毛片视频在线观看| 在线观看人妻少妇| 一级毛片 在线播放| 成年av动漫网址| 精品国产一区二区久久| 999久久久国产精品视频| 又大又黄又爽视频免费| 国产一级毛片在线| 欧美日韩一级在线毛片| 三上悠亚av全集在线观看| 成人黄色视频免费在线看| 免费看av在线观看网站| 国产精品成人在线| 久久久精品区二区三区| 精品人妻偷拍中文字幕| 亚洲欧美精品综合一区二区三区 | 国产精品人妻久久久影院| 国产亚洲精品第一综合不卡| 一级毛片 在线播放| 亚洲国产毛片av蜜桃av| 自拍欧美九色日韩亚洲蝌蚪91| 日韩电影二区| 2018国产大陆天天弄谢| 成年女人在线观看亚洲视频| 日韩av免费高清视频| 人妻 亚洲 视频| 国产人伦9x9x在线观看 | 欧美另类一区| 亚洲精华国产精华液的使用体验| 国产男女内射视频| 久久久欧美国产精品| 青春草亚洲视频在线观看| 18禁国产床啪视频网站| 青春草亚洲视频在线观看| 1024香蕉在线观看| 人人妻人人添人人爽欧美一区卜| 捣出白浆h1v1| 在线观看免费高清a一片| 99精国产麻豆久久婷婷| 亚洲欧美清纯卡通| 国产成人精品久久二区二区91 | 中文字幕最新亚洲高清| 人人妻人人爽人人添夜夜欢视频| 人人妻人人添人人爽欧美一区卜| 在线精品无人区一区二区三| av在线播放精品| 90打野战视频偷拍视频| 精品国产国语对白av| 国产成人精品久久久久久| 亚洲四区av| 最近最新中文字幕大全免费视频 | 免费在线观看黄色视频的| 精品亚洲乱码少妇综合久久| 色婷婷av一区二区三区视频| 水蜜桃什么品种好| 卡戴珊不雅视频在线播放| av有码第一页| 午夜91福利影院| 男女免费视频国产| 久久久久国产精品人妻一区二区| 午夜日本视频在线| 亚洲国产欧美日韩在线播放| 成人国产麻豆网| 三上悠亚av全集在线观看| 七月丁香在线播放| 国产免费又黄又爽又色| av一本久久久久| 久久久久精品人妻al黑| 在线观看免费高清a一片| 国产精品一区二区在线观看99| 又黄又粗又硬又大视频| 永久免费av网站大全| 美女国产高潮福利片在线看| 亚洲人成电影观看| 女的被弄到高潮叫床怎么办| 七月丁香在线播放| 日韩视频在线欧美|