• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Evolutionary GAN–Based Data Augmentation for Cardiac Magnetic Resonance Image

    2021-12-14 10:30:42YingFuMinxueGongGuangYangHongWeiandJiliuZhou
    Computers Materials&Continua 2021年7期

    Ying Fu,Minxue Gong,Guang Yang,Hong Wei and Jiliu Zhou,2

    1School of Computer Science,Chengdu University of Information Technology,Chengdu,610225,China

    2Images and Spatial Information 2011 Collaborative Innovation Center of Sichuan Province,Chengdu,610225,China

    3Department of Computer Science,University of Reading,Earley,RG6 6AY,UK

    Abstract: Generative adversarial networks (GANs) have considerable potential to alleviate challenges linked to data scarcity.Recent research has demonstrated the good performance of this method for data augmentation because GANs synthesize semantically meaningful data from standard signal distribution.The goal of this study was to solve the overfitting problem that is caused by the training process of convolution networks with a small dataset.In this context,we propose a data augmentation method based on an evolutionary generative adversarial network for cardiac magnetic resonance images to extend the training data.In our structure of the evolutionary GAN,the most optimal generator is chosen that considers the quality and diversity of generated images simultaneously from many generator mutations.Also,to expand the distribution of the whole training set,we combine the linear interpolation of eigenvectors to synthesize new training samples and synthesize related linear interpolation labels.This approach makes the discrete sample space become continuous and improves the smoothness between domains.The visual quality of the augmented cardiac magnetic resonance images is improved by our proposed method as shown by the data-augmented experiments.In addition, the effectiveness of our proposed method is verified by the classification experiments.The influence of the proportion of synthesized samples on the classification results of cardiac magnetic resonance images is also explored.

    Keywords: Evolutionary generative adversarial network; cardiac magnetic resonance; data augmentation; linear interpolation

    1 Introduction

    Cardiac magnetic resonance imaging (MRI) is the gold standard for assessing cardiac function.Conventional cardiac MRI scanning technology has advanced over the years and plays a vital role in the diagnosis of disease.Currently, many cardiac magnetic resonance imageassisted diagnosis tasks that are based on deep learning [1] have achieved good results.A novel algorithm [2] was proposed by Renugambal et al.for multilevel thresholding brain image segmentation in magnetic resonance image slices.This approach requires expensive medical equipment to obtain cardiac magnetic resonance images and also needs experienced radiologists to label them manually.It is undoubtedly extremely time-consuming and labor-intensive.The privacy of patients in the field of medical imaging has always been very sensitive and it is expensive to obtain a large number of datasets that are balanced between positive and negative samples.

    A significant challenge in the field of medical imaging based on deep learning is how to deal with small-scale datasets and a limited number of labeled data.Datasets are often not sufficient or the dataset sample is unbalanced, especially when using a complex deep learning model,which makes the deep convolution neural network with a huge number of parameters appear as overfitting [3].In the field of computer vision, scholars have proposed many effective methods for overfitting, such as batch regularization [4], dropout [5], early stopping method [6], weight sharing [7], weight attenuation [8], and others.These methods are to adjust the network structure.Data augmentation [9] is an effective method to operate on the data itself, which alleviates to a certain extent the problem of overfitting in image analysis and classification.The classical data augmentation techniques mainly include affine transformation methods such as image translation,rotation, scaling, flipping, and shearing [10,11].These approaches mix the original samples and new samples as training sets and input them into a convolutional neural network.The adjustment of the color space of samples is also a data augmentation method.Sang et al.[12] used the method of changing the brightness value to expand the sample size.These methods have improved solving the overfitting problem.However, the operation on the original samples does not produce new features.The diversity of the original samples has not been substantially increased [13],and the promotion effect is weak when processing small–scale data.Liu et al.[14] used a data augmentation method on the test set based on multiple cropping.Pan et al.[15] presented a novel image retrieval approach for small-and medium-scale food datasets, which both augments images utilizing image transformation techniques to enlarge the size of datasets and promotes the average accuracy of food recognition with deep learning technologies.

    The generative adversarial network (GAN) [16] is a generative model proposed by Ian Goodfellow and others.It consists of a generatorGand a discriminatorD.The generatorGuses noise z sampled from the uniform distribution or normal distribution as input to synthesize imageG(z).The discriminatorDattempts to judge the synthetic imageG(z)as false as much as possible and judge the real imagexas true.Next, it adjusts the parameters of each model through successive confrontation training.Finally, the generator obtains the distribution model of real samples and good performance of generating images close to real images.The specific structure of the GAN is shown in Fig.1.

    The entire GAN training process is designed to find the balance between the generative network and the discriminative network.This makes the discriminator unable to judge whether the samples generated by the generator are real so that the generative network can achieve the optimal performance.This process can be expressed as formula (1):

    The GAN generates new samples by fitting the original sample distribution.The new samples are generated from the distribution learned by the generative model, which makes it have new features that are different from the original samples.This characteristic makes it possible to use the samples generated by the generative network as new training samples to reach the goal of data expansion.The GAN has achieved good results in many computer vision fields.However,it has many problems in practical applications.It is very difficult to train a GAN.Once the data distribution and the distribution fitted by the generative network do not substantially overlap at the beginning of training, the gradient of the generative network can easily point to a random direction, which results in the problem of gradient disappearance [17].The generator will try to generate a single sample that is relatively conservative but lacks diversity to make the discriminator give high scores which leads to the problem of mode collapse [18].

    Figure 1:The structure of the GAN

    Many GAN variant models have been proposed to alleviate the situation of gradient disappearance and mode collapse.The more representative ones are deep convolutional GAN(DCGAN) [19], which combines a convolutional neural network with GAN, and conditional GAN [20] which adds a precondition control generator to the input data.The triple-GAN [21]adds a classifier based on the discriminator and generator, which can ensure that the classifier and generator achieve the optimal solution for classification from the perspective of game theory.In this approach, however, it is necessary to manually label samples.The improved Triple-GAN [22] method solves this problem and avoids the situation of gradient disappearance and training instability.In addition, the least squares GAN (LSGAN) [23] and wasserstein GAN(WGAN) [24] have made great improvements to the loss function.The WGAN uses wasserstein distance to measure the distribution distance, which makes the GAN training process more stable to a large extent.However, Gulrajan et al.[25] found that the WGAN uses a forced phase method to make the parameters of the network mostly focus on ?0.01, 0.01, which wastes the fitting ability of the convolutional neural network.Therefore, they proposed the WGAN-GP model,which effectively alleviated this problem.Li et al.[26] introduced the gradient penalty term to the WGAN network to improve the convergence efficiency.The evolutionary GAN proposed by Wang et al.[27] is a variant model of a generative adversarial network that is based on evolutionary algorithms.It will perform mutation operations when the discriminator stops training to generate multiple generators as adversarial targets and uses a specific evaluation method to evaluate the quality and diversity of the generated images in different environments (the current discriminator).This series of operations can reserve one or more generators with better performance for the next round of training.This method that overcomes the limitations of a single adversarial target is able to keep the best offspring all the time.It effectively alleviates the problem of mode collapse and improves the quality of the generator.

    Recently, many scholars have used GANs to augment training data samples.GANs have been used to augment the data of human faces and handwritten fonts [28].Ali et al.[29] used the improved PGGAN to expand the skin injury dataset and increased the classification accuracy.Frid et al.[30] used a DCGAN and an ACGAN to expand the data of liver medical images and proved that the DCGAN has a greater improvement in the classification effect on this dataset.In contrast to affine transformation, a GAN can be used to generate images with new features by learning the real distribution.

    The evolutionary GAN can improve the diversity and quality of generated samples.This study therefore uses an evolutionary GAN to perform data augmentation on cardiac magnetic resonance images.The main contributions of this study are as follows:

    (1) A cardiac magnetic resonance image data augmentation method based on an evolutionary GAN is proposed.This method generates high-quality and diverse samples to expand the training set and improves the value of various indicators of the classification results.

    (2) Linear interpolation of feature vectors is combined with the evolutionary GAN to synthesize new training samples and generate related linear interpolation labels.This not only expands the distribution of the entire training set, but also makes the discrete sample space continuous and improves the smoothness between domains, which better trains the model.

    (3) Various indicators of downstream classification tasks are used to optimize the model and the intensity of the experimental details.

    2 Evolutionary GAN

    The training process of the evolutionary GAN can be divided into three stages:mutation,evaluation, and selection.The first stage is mutation where the parent generator is mutated into multiple offspring generators; the second stage is evaluation where the adaptive score is worked out for each offspring generator of the current discriminator using an adaptive function; the third stage is selection where the offspring generator with the highest adaptive score is selected by sorting.The basic structure of the evolutionary GAN is shown in Fig.2.

    Figure 2:The structure of the evolutionary GAN

    2.1 Mutation

    The evolutionary GAN uses different mutation methods to obtain offspring generators based on parent generators.These mutation operators are different training targets, the purpose of which is to reduce the distance between the generated distribution and the real data distribution from different angles.It should be noted that the best discriminatorD* in formula (2) should be trained before each mutation operation.

    Zhang et al.[31] proposed three mutation methods:

    1) Maximum and minimum value mutation:Here the mutation has made little change to the original objective function, which provides an effective gradient and alleviates the situation of gradient disappearance.It can be written as formula (3):

    2) Heuristic mutation:Heuristic mutation aims to maximize the log probability of the error of the discriminator.The heuristic mutation will not be saturated when the discriminator judges the generated sample as false and it still provides an effective gradient to get the generator continuously trained.It can be written as formula (4):

    3) Least-squares mutation:Least-squares mutation can also avoid gradient disappearance.At the same time, compared with heuristic mutation, the least-square mutation does not generate false samples at a very high cost.It does not avoid punishment at a low cost, which can avoid mode collapse to a certain extent.It can be written as formula (5):

    2.2 Adaptive Function

    The evolutionary GAN uses the adaptive function to evaluate the performance of the generator and subsequently quantifies it as the corresponding adaptive score, which can be written as formula (6):

    Fqis used to measure the quality of the generated samples, namely whether the offspring generator can fool the discriminator, which can be written as formula (7):

    Fqmeasures the diversity of the generated samples.It measures the gradient generated when the parameters of the discriminator are updated again according to the offspring generator.If the samples generated by the offspring generator are relatively concentrated (lack of diversity), it is easier to cause a large fluctuation of the gradient when updating the discriminator parameters.This can be written as formula (8):

    γ(≥0)is a hyperparameter used to adjust the quality of samples generated and the weight of diversity, which can be adjusted freely in the experiment.

    3 Method

    In this study, we describe the design of a data augmentation model for cardiac magnetic resonance medical images based on an evolutionary GAN.This approach can generate highquality and diverse samples to expand the training set.The linear interpolation of related labels is generated by combining the linear interpolation of feature vector with the evolutionary GAN,which expands the distribution of the training set and makes the discrete sample space continuous to train the model better.The specific network structure is shown in Fig.3:

    Figure 3:The proposed network

    3.1 DAE GAN

    High quality and diversity of samples are needed when using a GAN for data augmentation.The evolutionary GAN is very suitable for data augmentation since it can be trained in a stable way and generates high-quality and diverse samples.The user can choose to focus on diversity or quality according to needs by adjusting the parameters in the adaptive function, which can make the process of data augmentation more operative.This study improves the evolutionary GAN and we name the improved model data augmentation evolutionary GAN (DAE GAN).

    There is no difference in the input and output between the evolutionary GAN and vanilla GAN.The only exception is that after fixing the discriminator parameters multiple offspring generators are mutated based on the parent generator for training.The optimal one or more generators is selected as the parent generator in the next discriminator environment, after the evaluation by the adaptive function.

    The evolutionary GAN greatly improves the diversity of generated samples.However, a certain number of training samples are required if we want to fully train the GAN model.In the case of too few training samples, the generator and discriminator are prone to reach the equilibrium point prematurely and also cause the problem of mode collapse in the generated data.This study uses the traditional affine transformation data augmentation methods before training the GAN to alleviate this problem, expanding the data by operations of horizontal flip, vertical inversion,translation, rotation, and others.The security of medical images was given careful consideration.We therefore did not add the original data with noise and avoided performing operations like cropping.In this way, we preserved the texture and edge features of the original data as far as possible.Traditional data augmentation only makes small changes to the original data and does not generate new features and the samples are also discrete.Thus, this study introduces linear interpolation.

    Zhang et al.[31] proposed a data augmentation method that is irrelevant to the data described in their article.This method constructs virtual training samples from original samples, combines linear interpolation of feature vectors to synthesize new training samples, and generates related linear interpolation labels to expand the distribution of the entire training set.The specific formula is as follows (9):

    xi,xjare the original input vectors;yi,yjare the label codes; (xi,xj) and (yi,yj) are two samples randomly sampled from the original samples;λ∈Beta[α,α] is the weight vector; andα∈(0,+∞)is the hyperparameter that controls the interpolation strength between the feature and the target vector.The linear interpolation method makes the model have the characteristics of a linear model.Processing the area between the original samples and the training samples reduces the inadaptability of predicting test samples other than the training samples.This enhances the generalization ability, simultaneously making the discrete sample space continuous and improving the smoothness between domains.

    The original input of the evolutionary GAN discriminator ought to be two samples after fixing the generator parameters:one is the generated sample (the discriminator tries to minimize the distance between the predicted label of this sample and “0”); the other is the real sample(the discriminator tries to minimize as much as possible the distance between the predicted label of this sample and “1”).The discriminator loss function of the original evolutionary GAN is described as follows in formula (10):

    The expanded expression can be written as formula (11):

    This study operates linear interpolation on the evolutionary GAN to modify the discriminator input from the original two pictures to one picture.The discriminator task is changed to minimize the distance between the predicted label of the fusion sample and ‘λ’.The loss function of the discriminator is modified as formula (12):

    3.2 Algorithm

    Typically, the GAN uses the noisezthat obeys the multivariate uniform distribution or multivariate normal distribution as the input of the model.Ben et al.[32] believed that multiple Gaussian distributions can better adapt to the inherent multimodal of the real training data distribution.They thus used multimodal distribution as an input in the GAN and demonstrated that this method can improve the quality and variety of generated images.The DAE GAN training process is shown in Tab.1.

    Table 1:The training process of the DAE GAN

    4 Results and Analysis

    4.1 Data Set and Preprocessing

    The magnetic resonance data in this experiment were obtained from a partner hospital.All samples are two-dimensional short-axis primary T1 mapping images.The spatial distance of these cardiac magnetic resonance images ranges from 1.172 × 1.172 × 1.0 mm3 to 1.406 × 1.406 ×1.0 mm3and the original pixel size is 256 × 218 × 1.The benign and malignant annotation and segmentation areas of these images are manually labeled and drawn by senior experts.The original image data is in the “.mha” for-mat.After a series of preprocessing operations, such as resampling, selection of regions of interest, normalization, and final selection of interest, we obtained a total of 298 images that consisted of 221 cardiomyopathy images and 77 non-diseased images.The size of the preprocessed image is 80 × 80 × 1.The preprocessed cardiac magnetic resonance image is shown in Fig.4.

    Figure 4:Cardiac magnetic resonance image region of interest

    All samples were normalized in this experiment to ensure the consistency of training data.Weused affine transformations on the training set before training the GAN.This included horizontal flip, vertical flip, 90°, 180°, 270°rotation, 0°–20°random rotation and amplification, 0%–2%random rescaling of the vertical and horizontal axes, the small and specific amplitude of rotation and amplification, and translation and amplification.The goal of these steps was not to lose the original image information from the data.After augmenting the training set once, we performed two kinds of operations on the data:the first was to put it into the classifier for training directly,and to subsequently use the test set to get the classification results; the second was to put it into different GANs for training and generating new samples to train the classifier again.

    4.2 Training the DAE GAN

    The original evolutionary GAN uses the structure of a DCGAN.In this study, we use the residual structure shown in Fig.5 in the generator and discriminator since the residual structure [33] can alleviate the gradient vanishing problem and accelerate the convergence rate.The goal is to train the high-performance generator more quickly in the same training time.

    After adding the self-attention module [34], the detailed structure of the generator and discriminator and the output size of each layer is shown in Tab.2.

    The DAE GAN experimental environment is as follows:Ubuntu 16.04.1 TLS, Tensorflow 1.14.0, two Nvidia Tesla M40 GPU with 12 GB video memory (used to train the generative models of diseased and non-diseased samples).The maximum storage capacity of the model is set to 4 to take into account the space occupation and accidental interruption.

    Figure 5:Residual block structure

    Table 2:The structure of the DAE GAN

    4.3 The Generation Results of the DAE GAN

    In this experiment, we use 5-fold cross-validation to dynamically divide the heart magnetic resonance image into a training set and a test set at a ratio of 0.8:0.2.We use only the training set in the DAE GAN training.Each model has been trained several times (≥5) in the experiment due to the uncertainty in the training process for the deep convolution model.The specific effect of the data augmentation method was verified by the average classification results.

    The training set of the cardiac magnetic resonance image data is expanded after normalization and affine transformation.We train the DAE GAN model by following the steps of Algorithm 1.The effects of our approach on the samples generated in the training process of the generative model are shown in Fig.6.

    Figure 6:The changing process of the generated sample

    A comparison of the samples generated by the trained generator versus the real samples is shown in Fig.7.

    4.4 Classification Experiment and Analysis of Experimental Results

    The observation method has strong subjectivity.In this experiment, data augmentation is performed on small sample medical images.Consequently, the observation method can only be used as a reference evaluation standard.Our study uses the ResNet50 model and the Xception model [35] as a classifier to evaluate the effect of data augmentation.The classification results are used to uniformly evaluate the effects of various data augmentation methods.

    In addition to the conventional accuracy index, we also calculate the two medical image classification indexes:sensitivity and specificity.These indicators are briefly explained here.

    The accuracy rate is the probability that the diseased sample and the non-diseased sample are judged correctly.The calculation formula is described as follows in formula (13):

    Sensitivity is the probability that a diseased sample is judged to be diseased.The calculation formula is described by formula (14):

    Figure 7:Comparison of generated samples with real samples.(a) Generated non-diseased samples(b) Generated non-diseased samples (c) Real non-diseased samples (d) Real diseased samples

    Specificity is the probability of judging a non-diseased sample as non-diseased.The calculation formula is described as in formula (15):

    TP stands for True Positive, which means that not only does the classifier judge it to be a diseased sample, but it is also a diseased sample.TN stands for True Negative.In this case, the classifier judges it to be a non-diseased sample, but it is in fact not a diseased sample.FP is short for False Positive, which means that the classifier judges it to be a diseased sample and it is a non-diseased sample.FN is short for False Negative.Here the classifier judges that the sample is not diseased but it is a diseased sample.

    In this study, we use the Keras framework under the Ubuntu 16.04.1 TLS system environment for the classification experiment (version number 2.24).We use a Tesla M40 in the training process.The learning rate is set to 1e?4, and we use the RMSprop optimizer.We set the early stopping method to prevent overfitting and the fivefold cross-validation method is used to find the average classification result of the classifier.The average classification results of each augmentation method in the ResNet50 and Xception classification models are shown in Tab.3.

    Table 3:The classification results of enhancement methods

    In the experiment, it is clear that the classification effect is not necessarily better as the number of generated samples increases.After adding a certain amount of data, the classification effect does not rise but decreases.At the same time, if only using generated samples without affine transformation data augmentation, the classification effect was not greatly improved compared with using affine transformation data augmentation alone.The specific experimental results are shown in Fig.8.

    Figure 8:The impact of data volume on classification results

    The experimental results show that we cannot completely obtain the original data distribution because the quality of the generated data remains poorer than the original data.The classification effect is slightly reduced by using only the generated data without the affine transformation data augmentation method.However, when the two methods are combined, the classification result of the classifier in-creases with the increase of the generated data and reaches the peak value when adding three times of the generated data.The addition of too much generated data leads to overfitting of the classification model and reduces the accuracy of classification.

    Experiments were performed to compare the different models with the classification results without any data augmentation method.For the ResNet50 model, the classification accuracy increased from 0.7767 to 0.8478, the sensitivity increased from 0.9674 to 0.9772, and the specificity increased from 0.6964 to 0.7822.For the Xception model, the classification accuracy increased from 0.7953 to 0.8698, the sensitivity increased from 0.9765 to 0.9798, and the specificity increased from 0.6833 to 0.8116.

    5 Conclusion

    The DAE GAN model proposed in this paper can effectively expand the amount of cardiac magnetic resonance image data, alleviating the problem of the classification network not being fully trained due to the small amount of medical image data and uneven data.The classification accuracy of the DAE GAN in ResNet50 and Xception models was increased by 7.11% and 7.45%, respectively, compared with not using data augmentation methods.The method proposed in this paper increased the classification accuracy in ResNet50 and Xception by 3.85% and 4.19%,respectively, compared with affine trans-formation data augmentation, and the experimental results showed that the method is effective in different classification models.

    Acknowledgement:We thank LetPub (www.letpub.com) for its linguistic assistance during the preparation of this manuscript.

    Funding Statement:Y.F.received funding in part from the Sichuan Science and Technology Program (http://kjt.sc.gov.cn/) under Grant 2019ZDZX0005 and the Chinese Scholarship Council(https://www.csc.edu.cn/) under Grant 201908515022.

    Conficts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲精品国产色婷婷电影| 久热爱精品视频在线9| 精品乱码久久久久久99久播| 无限看片的www在线观看| 久久久精品国产亚洲av高清涩受| 亚洲成人国产一区在线观看| 日韩精品中文字幕看吧| 桃色一区二区三区在线观看| 亚洲中文av在线| 午夜精品国产一区二区电影| 成人18禁高潮啪啪吃奶动态图| a级毛片在线看网站| 精品国产一区二区三区四区第35| 又黄又粗又硬又大视频| 可以免费在线观看a视频的电影网站| 俄罗斯特黄特色一大片| 欧美日本亚洲视频在线播放| 丝袜美足系列| 亚洲成国产人片在线观看| 免费人成视频x8x8入口观看| 亚洲片人在线观看| 国产亚洲精品综合一区在线观看 | 国产乱人伦免费视频| 色综合欧美亚洲国产小说| 高清黄色对白视频在线免费看| 日韩精品免费视频一区二区三区| 亚洲自偷自拍图片 自拍| 可以免费在线观看a视频的电影网站| 欧美日韩中文字幕国产精品一区二区三区 | 亚洲自偷自拍图片 自拍| 国产成人av激情在线播放| 欧美人与性动交α欧美精品济南到| 久久久精品欧美日韩精品| 免费在线观看完整版高清| 悠悠久久av| 亚洲欧美一区二区三区久久| 97超级碰碰碰精品色视频在线观看| 视频区欧美日本亚洲| 精品电影一区二区在线| 欧美乱码精品一区二区三区| 老司机午夜十八禁免费视频| 一个人免费在线观看的高清视频| 亚洲一区二区三区色噜噜 | 校园春色视频在线观看| 久久99一区二区三区| 午夜a级毛片| 一级黄色大片毛片| 国产av一区在线观看免费| 中文欧美无线码| 久久久水蜜桃国产精品网| 天天躁夜夜躁狠狠躁躁| 9191精品国产免费久久| 国产熟女午夜一区二区三区| 国产一区二区三区综合在线观看| 免费在线观看完整版高清| 色综合婷婷激情| 在线av久久热| 香蕉丝袜av| 亚洲欧美日韩无卡精品| 色婷婷久久久亚洲欧美| 精品欧美一区二区三区在线| 狠狠狠狠99中文字幕| 另类亚洲欧美激情| 一边摸一边做爽爽视频免费| 看免费av毛片| 久久久精品国产亚洲av高清涩受| 免费高清在线观看日韩| 国产精品亚洲一级av第二区| 夜夜躁狠狠躁天天躁| 精品熟女少妇八av免费久了| 国产精品爽爽va在线观看网站 | 久久久国产一区二区| 国产免费现黄频在线看| 欧美国产精品va在线观看不卡| 在线观看免费午夜福利视频| 五月开心婷婷网| 女人高潮潮喷娇喘18禁视频| 欧美国产精品va在线观看不卡| 国产高清videossex| 国产黄色免费在线视频| 啦啦啦 在线观看视频| 国产av在哪里看| 曰老女人黄片| av天堂久久9| 亚洲一区高清亚洲精品| 久久久久国内视频| 久久久精品欧美日韩精品| 国产精品二区激情视频| av中文乱码字幕在线| 亚洲一区二区三区色噜噜 | 免费在线观看亚洲国产| 免费搜索国产男女视频| 免费高清视频大片| 婷婷精品国产亚洲av在线| 欧美日韩精品网址| 高清在线国产一区| 亚洲欧洲精品一区二区精品久久久| 最近最新中文字幕大全免费视频| 国产成年人精品一区二区 | 香蕉丝袜av| 国产欧美日韩一区二区三| 午夜视频精品福利| av中文乱码字幕在线| 日韩 欧美 亚洲 中文字幕| 久久狼人影院| 天天影视国产精品| 久久精品aⅴ一区二区三区四区| 宅男免费午夜| 亚洲久久久国产精品| 国产一区二区激情短视频| 亚洲精品美女久久久久99蜜臀| 欧美激情极品国产一区二区三区| 欧美激情极品国产一区二区三区| 日本黄色日本黄色录像| 叶爱在线成人免费视频播放| 一夜夜www| a级毛片在线看网站| 国产熟女午夜一区二区三区| 午夜老司机福利片| 国产亚洲欧美98| 国产亚洲欧美98| 国产成人啪精品午夜网站| 久久伊人香网站| 久久久国产成人精品二区 | 身体一侧抽搐| 国产av在哪里看| 久久精品aⅴ一区二区三区四区| 亚洲成人免费av在线播放| 亚洲全国av大片| 99国产极品粉嫩在线观看| 每晚都被弄得嗷嗷叫到高潮| 欧美激情久久久久久爽电影 | 丰满的人妻完整版| 亚洲精品国产区一区二| 老司机亚洲免费影院| 精品国内亚洲2022精品成人| 久久久国产成人精品二区 | 国产精品久久视频播放| 91麻豆精品激情在线观看国产 | 成人亚洲精品一区在线观看| 国产一区在线观看成人免费| 亚洲五月色婷婷综合| 手机成人av网站| 丝袜在线中文字幕| 久久久久亚洲av毛片大全| 欧美日韩av久久| netflix在线观看网站| 怎么达到女性高潮| 国产高清国产精品国产三级| 午夜久久久在线观看| 国产精品偷伦视频观看了| 国产成人精品无人区| 90打野战视频偷拍视频| 国产av一区二区精品久久| 久久欧美精品欧美久久欧美| 国产精品1区2区在线观看.| 咕卡用的链子| 亚洲在线自拍视频| 女人被狂操c到高潮| 欧美日韩福利视频一区二区| 男人操女人黄网站| 男人的好看免费观看在线视频 | 久久香蕉精品热| 香蕉久久夜色| 看免费av毛片| 国产区一区二久久| 久久精品成人免费网站| av在线天堂中文字幕 | 999久久久国产精品视频| 超碰成人久久| 脱女人内裤的视频| 国产人伦9x9x在线观看| 午夜老司机福利片| 无人区码免费观看不卡| 99香蕉大伊视频| 嫁个100分男人电影在线观看| a级毛片在线看网站| 欧美在线一区亚洲| 一夜夜www| 亚洲视频免费观看视频| 男女床上黄色一级片免费看| 天堂动漫精品| 91精品三级在线观看| 国产成人精品在线电影| 国产成人精品无人区| 久久久久九九精品影院| 日韩有码中文字幕| 日本vs欧美在线观看视频| 欧美一级毛片孕妇| 亚洲情色 制服丝袜| 欧美日韩国产mv在线观看视频| 久久久久国产精品人妻aⅴ院| 亚洲三区欧美一区| 日韩欧美国产一区二区入口| 一a级毛片在线观看| 日本精品一区二区三区蜜桃| 国产亚洲欧美在线一区二区| 免费女性裸体啪啪无遮挡网站| 久久久久久久久久久久大奶| 欧美人与性动交α欧美精品济南到| 国产97色在线日韩免费| 亚洲欧美精品综合久久99| 国内久久婷婷六月综合欲色啪| 国产激情欧美一区二区| 一边摸一边抽搐一进一出视频| 久久久久久人人人人人| 国产精品香港三级国产av潘金莲| 亚洲中文av在线| av超薄肉色丝袜交足视频| 欧洲精品卡2卡3卡4卡5卡区| 在线播放国产精品三级| 国内久久婷婷六月综合欲色啪| 亚洲欧美一区二区三区久久| 久久久久久久久中文| 在线观看免费视频日本深夜| 国产无遮挡羞羞视频在线观看| 亚洲av熟女| 欧美日韩乱码在线| 91av网站免费观看| 亚洲中文av在线| 亚洲精品美女久久久久99蜜臀| 99热只有精品国产| 亚洲,欧美精品.| 88av欧美| 国产精品国产av在线观看| av欧美777| 精品国产一区二区三区四区第35| 精品一品国产午夜福利视频| 亚洲久久久国产精品| 亚洲激情在线av| 精品一品国产午夜福利视频| 国产免费现黄频在线看| 视频区欧美日本亚洲| 亚洲熟妇熟女久久| 亚洲精品国产精品久久久不卡| 亚洲精品国产精品久久久不卡| 99在线人妻在线中文字幕| 人人妻人人爽人人添夜夜欢视频| 他把我摸到了高潮在线观看| 婷婷六月久久综合丁香| 十八禁网站免费在线| 成人亚洲精品一区在线观看| 亚洲狠狠婷婷综合久久图片| 99精国产麻豆久久婷婷| 日韩欧美免费精品| 天堂中文最新版在线下载| 欧美日韩亚洲综合一区二区三区_| 免费观看人在逋| 99国产综合亚洲精品| 欧美日韩乱码在线| 国产成人精品久久二区二区91| 满18在线观看网站| 久久精品成人免费网站| 99riav亚洲国产免费| 一个人免费在线观看的高清视频| 亚洲九九香蕉| 丝袜人妻中文字幕| 在线观看免费午夜福利视频| 正在播放国产对白刺激| 777久久人妻少妇嫩草av网站| 亚洲自拍偷在线| 91成年电影在线观看| 久久中文字幕人妻熟女| 在线观看免费午夜福利视频| √禁漫天堂资源中文www| 亚洲欧美激情综合另类| 老司机午夜福利在线观看视频| 亚洲欧美精品综合久久99| 韩国av一区二区三区四区| 欧美乱码精品一区二区三区| x7x7x7水蜜桃| 亚洲男人的天堂狠狠| 婷婷六月久久综合丁香| 悠悠久久av| 久久久久精品国产欧美久久久| 日本wwww免费看| 午夜福利一区二区在线看| 色老头精品视频在线观看| 中文字幕精品免费在线观看视频| 国产精品一区二区免费欧美| aaaaa片日本免费| 青草久久国产| 如日韩欧美国产精品一区二区三区| 18美女黄网站色大片免费观看| 1024视频免费在线观看| 国产精品 国内视频| 18美女黄网站色大片免费观看| 中亚洲国语对白在线视频| 精品国产一区二区三区四区第35| 80岁老熟妇乱子伦牲交| 少妇裸体淫交视频免费看高清 | 国产精品乱码一区二三区的特点 | 亚洲熟女毛片儿| 成人av一区二区三区在线看| 精品久久久久久久久久免费视频 | 交换朋友夫妻互换小说| 久久久久久大精品| 日本 av在线| 在线视频色国产色| avwww免费| 久久中文字幕一级| 少妇的丰满在线观看| 久久人妻av系列| 亚洲中文字幕日韩| 国产深夜福利视频在线观看| 亚洲成av片中文字幕在线观看| 亚洲国产精品sss在线观看 | 日本wwww免费看| 成年女人毛片免费观看观看9| 亚洲精品久久午夜乱码| 久久久久久人人人人人| 亚洲成人久久性| 日本黄色视频三级网站网址| 成人三级黄色视频| 午夜免费观看网址| 亚洲欧美一区二区三区黑人| 大码成人一级视频| 91在线观看av| 亚洲av成人不卡在线观看播放网| 亚洲片人在线观看| 亚洲av熟女| 亚洲午夜精品一区,二区,三区| 成年版毛片免费区| 丝袜在线中文字幕| 亚洲片人在线观看| 老鸭窝网址在线观看| 成年女人毛片免费观看观看9| 正在播放国产对白刺激| 精品福利观看| 国产精品香港三级国产av潘金莲| 19禁男女啪啪无遮挡网站| 亚洲国产欧美一区二区综合| 88av欧美| 久久久久国内视频| 久久性视频一级片| 成人亚洲精品av一区二区 | 亚洲欧美日韩高清在线视频| 咕卡用的链子| 国内久久婷婷六月综合欲色啪| av网站在线播放免费| 美女扒开内裤让男人捅视频| 9191精品国产免费久久| 老汉色∧v一级毛片| avwww免费| 久久中文字幕一级| 午夜福利在线观看吧| 国产麻豆69| 在线播放国产精品三级| 久久精品国产综合久久久| 又紧又爽又黄一区二区| 大型黄色视频在线免费观看| 在线天堂中文资源库| 99香蕉大伊视频| 中文字幕色久视频| 丁香六月欧美| 欧美日韩国产mv在线观看视频| ponron亚洲| 欧美一区二区精品小视频在线| 国产欧美日韩一区二区三| 在线看a的网站| 夜夜夜夜夜久久久久| 精品国产乱子伦一区二区三区| 美国免费a级毛片| 日本免费一区二区三区高清不卡 | 中文字幕人妻丝袜一区二区| 大型av网站在线播放| 久久香蕉激情| 中文字幕高清在线视频| 99久久人妻综合| 少妇被粗大的猛进出69影院| 淫秽高清视频在线观看| 久久精品亚洲av国产电影网| 级片在线观看| 色老头精品视频在线观看| 激情在线观看视频在线高清| 在线观看66精品国产| 黑人巨大精品欧美一区二区蜜桃| 国产亚洲欧美在线一区二区| 一区二区三区精品91| 丰满人妻熟妇乱又伦精品不卡| 老司机亚洲免费影院| 丰满饥渴人妻一区二区三| 成人av一区二区三区在线看| 成人国产一区最新在线观看| 国产蜜桃级精品一区二区三区| 嫩草影视91久久| 一级毛片女人18水好多| 一本大道久久a久久精品| 久久精品亚洲熟妇少妇任你| 12—13女人毛片做爰片一| 欧美老熟妇乱子伦牲交| 免费看十八禁软件| 午夜精品在线福利| 看免费av毛片| 激情视频va一区二区三区| 在线观看免费视频网站a站| 国产成年人精品一区二区 | 久久亚洲真实| 亚洲第一欧美日韩一区二区三区| 好看av亚洲va欧美ⅴa在| 国产精品综合久久久久久久免费 | 国产在线精品亚洲第一网站| 一本大道久久a久久精品| 一进一出抽搐gif免费好疼 | av有码第一页| 国产精品九九99| 最近最新免费中文字幕在线| 亚洲第一欧美日韩一区二区三区| 精品国产乱子伦一区二区三区| 欧美成狂野欧美在线观看| 亚洲欧美激情综合另类| 日韩欧美在线二视频| 日韩国内少妇激情av| 在线观看一区二区三区| 高清黄色对白视频在线免费看| 亚洲免费av在线视频| 免费在线观看黄色视频的| 国产成人免费无遮挡视频| 亚洲中文字幕日韩| 老司机靠b影院| 99re在线观看精品视频| 中文字幕人妻丝袜制服| 国产成人av教育| 黑人欧美特级aaaaaa片| 91九色精品人成在线观看| 黄网站色视频无遮挡免费观看| 日韩精品免费视频一区二区三区| 日韩欧美一区视频在线观看| 亚洲狠狠婷婷综合久久图片| 欧美日韩视频精品一区| 一级作爱视频免费观看| 欧美中文日本在线观看视频| av网站在线播放免费| 久久精品国产亚洲av香蕉五月| 一二三四社区在线视频社区8| 日韩人妻精品一区2区三区| e午夜精品久久久久久久| 99国产精品99久久久久| 一进一出抽搐gif免费好疼 | 午夜激情av网站| 老司机靠b影院| 欧美黑人欧美精品刺激| 嫁个100分男人电影在线观看| 性少妇av在线| 亚洲人成伊人成综合网2020| 亚洲欧美日韩高清在线视频| 欧美日韩瑟瑟在线播放| av天堂在线播放| 激情视频va一区二区三区| 国产精品久久久久成人av| 不卡一级毛片| 美女大奶头视频| 90打野战视频偷拍视频| 中亚洲国语对白在线视频| www.精华液| 妹子高潮喷水视频| 久久中文看片网| 啪啪无遮挡十八禁网站| 亚洲欧洲精品一区二区精品久久久| 久久婷婷成人综合色麻豆| 国产亚洲精品久久久久5区| 亚洲avbb在线观看| 亚洲三区欧美一区| 日本黄色日本黄色录像| 夜夜夜夜夜久久久久| 一边摸一边抽搐一进一出视频| 黄频高清免费视频| 在线播放国产精品三级| 99国产精品一区二区三区| 国产亚洲欧美在线一区二区| 在线播放国产精品三级| 日韩 欧美 亚洲 中文字幕| 国产亚洲精品久久久久5区| 久久久国产一区二区| 在线十欧美十亚洲十日本专区| 亚洲av成人av| 99久久精品国产亚洲精品| 久久精品国产综合久久久| 亚洲男人天堂网一区| 99久久国产精品久久久| 激情在线观看视频在线高清| 国产精品爽爽va在线观看网站 | 1024香蕉在线观看| 50天的宝宝边吃奶边哭怎么回事| a级片在线免费高清观看视频| 久久久国产成人精品二区 | 女生性感内裤真人,穿戴方法视频| 韩国av一区二区三区四区| 国产一区二区激情短视频| 精品第一国产精品| 免费av毛片视频| 脱女人内裤的视频| 亚洲男人天堂网一区| 丰满迷人的少妇在线观看| 亚洲精品久久成人aⅴ小说| 黄片播放在线免费| 国产欧美日韩精品亚洲av| 制服人妻中文乱码| 黄色视频不卡| 人人澡人人妻人| 婷婷六月久久综合丁香| 欧美最黄视频在线播放免费 | 欧美成狂野欧美在线观看| 国产视频一区二区在线看| 狠狠狠狠99中文字幕| 老司机亚洲免费影院| а√天堂www在线а√下载| 美女国产高潮福利片在线看| 电影成人av| 曰老女人黄片| 午夜亚洲福利在线播放| 久久香蕉精品热| 麻豆av在线久日| 777久久人妻少妇嫩草av网站| 这个男人来自地球电影免费观看| 亚洲成人免费电影在线观看| 18美女黄网站色大片免费观看| 亚洲自偷自拍图片 自拍| 欧美老熟妇乱子伦牲交| 国产片内射在线| 国产精品综合久久久久久久免费 | 欧美人与性动交α欧美软件| 在线看a的网站| 超碰成人久久| 国产亚洲精品久久久久5区| 久久国产亚洲av麻豆专区| 亚洲黑人精品在线| 欧美不卡视频在线免费观看 | 大香蕉久久成人网| 精品卡一卡二卡四卡免费| 美女高潮到喷水免费观看| 欧美一级毛片孕妇| 激情视频va一区二区三区| 99国产综合亚洲精品| 国产国语露脸激情在线看| 欧美久久黑人一区二区| 又紧又爽又黄一区二区| videosex国产| 老司机午夜福利在线观看视频| 777久久人妻少妇嫩草av网站| 国产精品二区激情视频| 夜夜夜夜夜久久久久| 女同久久另类99精品国产91| 19禁男女啪啪无遮挡网站| 老司机午夜福利在线观看视频| 人人妻人人澡人人看| 侵犯人妻中文字幕一二三四区| 午夜福利在线免费观看网站| 久久热在线av| 亚洲国产毛片av蜜桃av| 午夜亚洲福利在线播放| 人人澡人人妻人| 国产在线观看jvid| 嫁个100分男人电影在线观看| 看片在线看免费视频| 国产真人三级小视频在线观看| 性色av乱码一区二区三区2| 男女高潮啪啪啪动态图| 水蜜桃什么品种好| 欧美日本中文国产一区发布| 成人亚洲精品av一区二区 | 久热爱精品视频在线9| 露出奶头的视频| 人妻丰满熟妇av一区二区三区| 日日干狠狠操夜夜爽| 亚洲国产精品合色在线| 免费女性裸体啪啪无遮挡网站| 国产99白浆流出| 国产熟女午夜一区二区三区| 午夜视频精品福利| √禁漫天堂资源中文www| 国产欧美日韩一区二区三| 亚洲成国产人片在线观看| 午夜免费观看网址| www.熟女人妻精品国产| 在线观看一区二区三区激情| 免费观看精品视频网站| 久久精品影院6| 黄色毛片三级朝国网站| 中国美女看黄片| 亚洲成人免费电影在线观看| tocl精华| 亚洲片人在线观看| 正在播放国产对白刺激| 美女高潮喷水抽搐中文字幕| 不卡av一区二区三区| 国产成人精品无人区| 国产1区2区3区精品| 亚洲精华国产精华精| 国产精品一区二区精品视频观看| 久久久久久久久久久久大奶| 一区二区三区激情视频| 欧美成狂野欧美在线观看| 亚洲欧美日韩无卡精品| 美女扒开内裤让男人捅视频| 男女午夜视频在线观看| 真人一进一出gif抽搐免费| 黑人操中国人逼视频| 美女午夜性视频免费| 精品久久久久久成人av| 日韩国内少妇激情av| 日本精品一区二区三区蜜桃| 欧美乱妇无乱码| 欧美日韩瑟瑟在线播放| 天堂√8在线中文| 久久久久精品国产欧美久久久| 一区二区日韩欧美中文字幕| 国产人伦9x9x在线观看| 精品一品国产午夜福利视频| 在线观看舔阴道视频| 男女之事视频高清在线观看| 一本大道久久a久久精品| 国产极品粉嫩免费观看在线| 亚洲成人免费av在线播放| 久久青草综合色|