• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Image Color Rendering Based on Hinge-Cross-Entropy GAN in Internet of Medical Things

    2023-03-12 09:01:30HonganLiMinZhangDufengChenJingZhangMengYangandZhanliLi

    Hong’an Li,Min Zhang,★,Dufeng Chen,Jing Zhang,Meng Yang and Zhanli Li

    1College of Computer Science and Technology,Xi’an University of Science and Technology,Xi’an,710054,China

    2Beijing Geotechnical and Investigation Engineering Insititute,Beijing,100080,China

    3Xi’an Institute of Applied Optics,Xi’an,710065,China

    ABSTRACT Computer-aided diagnosis based on image color rendering promotes medical image analysis and doctor-patient communication by highlighting important information of medical diagnosis.To overcome the limitations of the color rendering method based on deep learning, such as poor model stability, poor rendering quality, fuzzy boundaries and crossed color boundaries,we propose a novel hinge-cross-entropy generative adversarial network(HCEGAN).The self-attention mechanism was added and improved to focus on the important information of the image.And the hinge-cross-entropy loss function was used to stabilize the training process of GAN models.In this study,we implement the HCEGAN model for image color rendering based on DIV2 K and COCO datasets,and evaluate the results using SSIM and PSNR.The experimental results show that the proposed HCEGAN automatically re-renders images, significantly improves the quality of color rendering and greatly improves the stability of prior GAN models.

    KEYWORDS Internet of Medical Things;medical image analysis;image color rendering;loss function;self-attention;generative adversarial networks

    1 Introduction

    The Internet of Things (IoT) is getting more popular and has a high level of interest in both practitioners and academicians in the age of wireless communication due to its diverse applications.And the significant increase in the number of individuals with chronic ailments has dictated an urgent need for an innovative model for healthcare systems[1–3].So IoT based on deep neural network model has been applied in medical treatment in recent years.The Internet of Medical Things(IoMT)based on computer-aided diagnosis is gradually developing faster and more efficiently [4–7].Image color rendering based on deep learning can render the grayscale image into a color image to highlight the deep information and help readers to quickly understand the depth of information contained in the image[8–11].With the rapid development of computer graphics,computer vision and related software and hardware technologies,the method of rendering existing images and enhancing details by using high-quality image data has attracted extensive research attention,and has been gradually applied to medical image processing and analysis in IoMT[12–14].

    At present, the traditional color rendering method needs manual intervention and requires high reference image.The color rendering method based on deep learning uses the neural network model and the corresponding high-quality image dataset to train the model, and the image can be automatically rendered according to the model without being affected by human factors or other factors.

    Image color rendering based on deep learning is generally divided into color rendering based on convolutional neural network and color rendering based on generative adversarial network [15,16].Iizuka et al.[17] used a fusion layer in a convolutional neural network to combine low-dimensional features and global features of an image to generate the image color.Zhang et al.[18] designed an appropriate loss function to deal with multi-mode uncertainty in color rendering and maintain diversity of colors.Sangkloy et al.[19] combined a method based on graffiti with deep learning and trained images with color lines in a neural network.He et al.[20] selectively migrated reference image colors that were consistent with a target image in semantic structure and content to the target image,otherwise learn the color from large-scale data.Xiao et al.[21]proposed a depth sample-based image rendering method using a dense coded pyramid[22–25].

    However, when extracting grayscale image features, the aforementioned method adopts upsampling to make the image size consistent,resulting in a loss of image information.Therefore,the network structure thus constructed cannot effectively extract and recognize complex features of the image.Hence,the rendering effect is relatively limited.Goodfellow et al.[26]proposed a generative adversarial network(GAN).GAN model based on unsupervised learning can approximate arbitrary distribution,so it has a wide range of applications in the field of image generation.But it is subject to the wellknown disappearing and exploding gradient problem due to the instability of the model,resulting in a deterioration of image rendering performance.On this basis,Mirza et al.[27]proposed a conditional generative adversarial network(CGAN),which is conditional on additional information,such as class tags or data from other modes,to enter the discriminant and generator as an additional input layer for adjustment.Isola et al.[28]improved a CGAN to transform image styles,for example,from grayscale to color images,from day to night images,from lines to shaded images.The proposed pix2pix model has a powerful image conversion function,and can learn a mapping relationship between a gray and color image to achieve color rendering.

    Although the GAN-based image color rendering method can automatically render images,it has certain problems such as blurred boundaries and unclear details.Moreover, unstable GAN models lead to low rendering quality[29].Therefore,Arjovsky et al.[30]used Wasserstein distance to replace the original JS divergence to stabilize the training of GANs, whereas Gulrajani et al.[31] added penalty terms to make the training of GANs more stable.Mao et al.[32] used the least square loss function to solve the problem of gradient disappearance and enhance the stability of the model.However,the stability of these GAN models and the effectiveness of the color rendering method still need to be improved.Therefore, this paper proposes a hinge-cross-entropy GAN (HCEGAN) for automatic image rendering.First,a new hinge-cross-entropy loss function is proposed to stabilize the GAN training.Second, an improved self-attention module is added to the GAN model to improve color rendering quality more quickly and effectively.Finally,by adding a Skip Connection,a network structure based on U-Net,as a generator for the GAN model,we adopt the 70×70 PatchGAN based on image conversion as a discriminator.Experimental results show that the HCEGAN has a more significant rendering effect compared with the pix2pix model.

    2 Related Works

    2.1 Hinge Loss

    The hinge loss allows the distance between incorrectly classified samples and correctly classified samples to be sufficiently far.Unless the difference between a threshold Δ,and an incorrect classification error is considered to be 0,a calculation error is accumulated[33].The loss function does not simply require the highest score in the right category,but rather a certain amount.Even a calculation classified as a loss may be correct,because it is possible that a correct category score will fail to exceed a certain threshold Δ.

    Suppose we classify some inputxi,whose label isyi,and predictxithrough a functionf(xi,W).Threshold Δ generally takes a value of 1.The output iss.Then,we predict the inputxias thejth type.Therefore,the output issj=f(xi,W)j.LetLirepresent the loss of each class,the loss of a sample is the sum of the losses of all classes.Then the calculation of the loss function of the outputxiis shown in Fig.1.

    Figure 1:Hinge loss

    Hinge loss is often used in binary classification problems.For outputy= ±1,threshold Δ takes the value of 1.Then,we predict the hinge loss of ?yas

    2.2 Cross-Entropy Loss

    Cross-entropy is used to measure the difference between two probability distributions in information theory.The cross-entropy loss function is not only simple and effective, but can also avoid the reduction in learning rate caused by the mean square error loss function in gradient decline using the sigmoid function[34].Assuming that the probability of an event isp,the amount of informationIis expressed as

    Entropy represents the measurement of the uncertainty of random variables.Under the discrete condition,entropy of information can be expressed as

    In the discrete condition,if there are two probability distributionsQandPfor the same random variablex,we can use KL divergence to measure the difference between these two distributions.

    The discrete distribution is the same,hence,the KL divergence is zero.Therefore,the cross-entropy is expressed as the sum of entropy and KL divergence.

    Then,in the case of binary classification,the cross-entropy loss of batch samples is

    At present, the cross-entropy loss is mainly used to classify the problem, combined with the Sigmoid or Softmax activation function.The advantage of cross-entropy loss is that the learning speed is fast when the model effect is poor, and slow when the model effect is good.However,because the cross-entropy loss combined with the Sigmoid or Softmax activation function adopts the inter-class competition mechanism, the features learned are scattered and multiple features cannot be distinguished.Therefore, the activation function can only be continuously optimized to better strengthen the model.

    3 Hinge-Cross-Entropy GAN

    3.1 Hinge-Cross-Entropy Loss

    The loss function in deep learning is mainly used to evaluate the degree to which the predicted sample is different from the real sample.In general,the better the loss function,the better the model performance.The original GAN model scales the output of the discriminator neural network to probability [0, 1] by the Sigmoid function, and measures the cross-entropy loss of probability.In order to minimize JS(Jensen-Shannon)difference between model distribution and target distribution,the minimax game of the GAN model is realized.Many scholars use regularizer or loss function to minimize the difference between model distribution and target distribution, and ensure the stability of the GAN model.As can be seen from the analysis of the two loss functions in Section 2, the hinge loss function is able to keep the distance between the incorrectly classified samples and the correctly classified samples sufficiently far beyond a certain threshold so that the unclassified error value remains 0.In the cross-entropy loss function,the closer the predicted output is to the real sample,the smaller the loss function is,and the closer the predicted function is to 1.In this case,the variation trend of the function is completely in line with the actual needs.Therefore,the greater the difference between the predicted output and the real sample, the greater the loss value, that is, the greater the penalty to the current model,according to a non-linear increase similar to exponential growth.

    This situation is mainly determined by the log function, which influences the model to tend to make the predicted output closer to the real sample.Further, the hinge loss optimization to enforce the requirements of remaining within less than a certain distance will cease optimization, whereas the cross-entropy loss is always optimized.Hence, in general, the cross-entropy loss is better than hinge loss.However,the cross-entropy loss is good at learning information between classes.Because it adopts the inter-class competition mechanism,the model will try to learn different types of features.And only consider the accuracy of the prediction probability of the correct samples,while ignoring the differences of other wrong samples.So the learned features are scattered,and the effect generated by the generator is suboptimal.Based on the advantages and disadvantages of hinge and cross-entropy loss,we propose a hinge-cross-entropy loss function.

    First,we define the loss function based on the pix2pix network model as follows:

    where,xis the input image,yis the expected output,Gis the generator, andDis the discriminator.In addition to the generating adversarial loss of CGAN model, the pix2pix model also adds a L1 regularization loss times a certain parameter to improve the training of the model.Andλis this parameter.

    Meanwhile,the hinged version of adversarial loss in GAN model is defined.We set the generated image asG(z).Then,the loss functions of the generator and discriminator are,respectively:

    On this basis,the hinge-cross-entropy loss function is defined as follows:

    The loss function of the generator is the cross-entropy loss function adding L1 loss,which is the same as the loss function of the generator of the pix2pix model.The loss function of the discriminant is in the form of the cross-entropy loss function after the real or generated image is processed by the hinge loss function.In this paper, the binary cross-entropy loss (BCE Loss) function is used, which is a special case of the cross-entropy loss function used for binary classification problems.The exact definition of the function is the same.Because this is a binary classification task,there are only two possibilities,plus or minus,and the sum of probabilities is 1,so we only have to predict one probability.In practical applications,a sigmoid function should be added to the layer of BCE Loss to normalize the data first before binary cross entropy loss can be used for calculation.The hinge-cross-entropy loss function also does not need to optimize the activation function to improve its problems such as large computation and small feature discrimination due to the large number of classification datasets.

    3.2 Improved Self-Attention Module

    The function of the attention mechanis m in computer vision is to enable a neural network model to learn to ignore irrelevant information and focus on the important information[35].Usually,relevant features are used to learn the weight distribution.Then,the learned weight is applied to the features to extract relevant knowledge further.The weight can be applied to the original image,or to the spatial scale, channel scale, etc.The self-attention mechanism is a concept learned from natural language processing(NLP).Other tasks in the direction of computer vision have also achieved good results.

    The basic structure of self-attention module is shown in Fig.2.It can be seen that the self-attention module is divided into three branches by a 1×1 convolution:f,gandh.First,the dot product is used to calculate the similarity betweenfandgto get the weight.Second, a softmax function is used to normalize these weights.Finally, the weighted sum of weights and correspondinghis calculated to get the final attention value.However,this weight parameter needs to be initialized to 0,which first depends on the local originalx, then gradually increases the non-local weight, and finally applies self-attention to the generator and discriminator.If the initial weight is set to 0 and the weight is gradually increased through model optimization, the generator will learn slowly, leading to slow iteration.Therefore, we need to adjust the learning method, especially the initialization matrix, to speed up the learning speed of the attention mechanism.

    Figure 2:Self-attention module

    In this paper,we adjust the initial weight to a diagonal matrix,and assume that the values on the diagonal of the diagonal matrix conform to the standard normal distribution with a mean of 0 and a variance of 1,such as the diagonal matrix A of size 3×3:

    The calculation steps of self-attention mechanism are as follows.First,we aim to transform the feature map of the previous hidden layerx∈RC×Ninto the feature spacef,gand calculate attention,wheref(x)=Wfx,g(x)=Wgx.Each elementβjiin the attention map represents how much the model pays attention to theith pixel when synthesizing thejth pixel,wheresij=f(xi)Tg(xj).

    Then, we useβjias the weight of attention to calculate the outputo=(o1,o2,...,oj,...,oN)∈RC×N,where,oj=βj,ih(xi),h(xi)=Whxi.Wf,Wg,andWhare the weight matrices learned after a 1×1 convolution.Here,Wf∈RC×C,Wg∈RC×C,Wh∈RC×C.

    Finally,we multiply the output of the attention layer by a scale parameterγand add it back to the input feature map.

    3.3 Hinge-Cross-Entropy GAN

    In this paper,we construct the generator of hinge-cross-entropy GAN based on U-Net structure,coding-decoding structure and Skip Connection, as shown in Fig.3.The orange module is the improved self-attention module,and the blue module is the convolution layer,which contains 8 convolution layers and 8 deconvolution layers.Lines represent Skip Connection.The U-Net structure is a symmetric U-shaped structure of compression path and expansion path.The parameter transfer and error feedback of deep neural network can be strengthened by adding Skip Connection.In addition,in order to solve the problem of remote dependence and strengthen long-distance information, an improved self-attention module is added in the encoder,namely,in front of each convolutional layer,respectively, in order to effectively select feature information.The input of neural network is many vectors of different sizes,and there are certain relationships among them.However,the relationship between these inputs can not be fully developed in the actual training,resulting in poor results.The selfattention mechanism can establish correlations between multiple related inputs,allowing the network model to notice correlations between different parts of the entire input.

    Figure 3:Proposed generator structure

    The discriminator of hinge-cross-entropy GAN uses 70 × 70 PatchGAN, which contains 4 convolution layers,each of which uses the unit form of conversion-regularization-Leakyrelu activation function.Instead of measuring the whole image with a single value, PatchGAN used an N * N matrix to evaluate the whole image, so that more areas could be focused on.The input/output size is 256×256,the step size is 2,the fill pixel is 1,the activation function is LeakyReLU,and BatchNorm is used.The whole GAN model adopts minimax game, generate color image as close as possible to the expected image,deceive the discriminator.Grayscale image and generated image are input into the discriminator at the same time.The discriminator tries to distinguish the generated image and see the difference between the generated image and the expected image.During model learning and training,generator and discriminator are trained alternately.After the training is complete,the generator is used to generate the desired color image.

    4 Experiments

    The DIV2 K dataset covers a wide range of contexts, including people, handmade objects and cities and villages, for example based on single-image super-resolution benchmarking.The dataset participating in the NTIRE 2017 Challenge drives the state-of-the-art technology in terms of singleimage super-resolution.There are 800 training sets and 100 testing sets in the DIV2 K dataset.The COCO2017 dataset is an advanced version of the Microsoft COCO dataset funded by Microsoft in 2014.Among them, the COCO competition was one of the most concerned and authoritative competitions in the field of computer vision at that time.The COCO datasets include six categories:fish, ladybug, orange, lion, dog and bird.There are 500 training sets and 100 testing sets for each category of the COCO dataset.

    When DIV2 K dataset was used to verify the validity of the model, the experimental environment used was Windows 10, with a 64-bit operating system, an Intel(R) Core(TM) I7–9750H CPU@2.5 GHz on a notebook computer,as well as Python 3.7,Pytorch1.2,and CUDA 10.0.When the COCO dataset was used, the experimental environment used was Windows 10, with a 64-bit operating system,an Intel(R)Core(TM)I9–10900x CPU@3.70 GHz on a desktop computer,as well as Python 3.7, Pytorch1.7, and CUDA 10.2.In experiments, the same parameters were used for all models under the PyTorch framework,the iteration was 200,the optimizer was Adam,and the learning rate was 0.0002.In order to reflect the color rendering quality of different models, the experiment adopts PSNR and SSIM to evaluate the rendered images.

    4.1 Qualitative Comparison

    Effect of self-attention moduleTo verify the effectiveness of the self-attention module,the added module was compared with the original pix2pix model,and the results on the DIV2 K dataset were shown in Figs.4b, 4c.By comparing the rendering results of the original color image, the pix2pix model and the model with self-attention module, it can be seen that the pix2pix model has a large error in rendering due to the instability of GAN model,resulting in color pollution in the upper left corner of Fig.4b.However,the model with self-attention module restores the real color of the image in both structure and color,and the overall tone is more harmonious.After the self-attention module is added,the self-attention mechanism can learn important features and suppress non-important features to achieve fast learning of image information.The results are shown in Table 1.After adding the self-attention module,SSIM and PSNR in the DIV2 K dataset were increased by 0.35%and 1.2 dB,respectively.In the COCO dataset,SSIM and PSNR were almost all improved.Therefore,after the selfattention module is added to the model,the module can improve the model’s attention to important information by using the inherent information of features for attention interaction,thus significantly improving the rendering quality.Therefore, adding self-attention module into the model is effective and improves the rendering effect of the model.

    Figure 4:Qualitative comparison

    Table 1: Verification of the effectiveness of the improved self-attention module

    Effect of hinge-cross-entropy lossTo verify the effectiveness of the hinge-cross-entropy loss function,we added the hinge-cross-entropy loss to the pix2pix model and compared it with the original pix2pix model by adding the mean-square error (MSE) loss.The results on the DIV2 K dataset are shown in Figs.4d, 4e.Comparing the rendering results of the original color image with the hingecross-entropy loss function and the mean square error loss function,it may be observed that the hingecross-entropy loss function has a better rendering result on the details of an image such as steps and roofs, and there is no color pollution.This is because the hinge-cross-entropy loss function inherits the advantages of the hing loss function and the cross-entropy loss function.This puts the distance between misclassified samples and correctly classified samples far beyond a certain threshold.And when the optimization of the hinge loss function is forced to keep within a certain distance,the crossentropy loss always keeps the optimal state.

    Similarly,we used PSNR and SSIM to evaluate the rendered images.The experimental results are shown in Table 2.Hinge stands for hinge loss function,BCE stands for cross-entropy loss function,and MSE stands for mean square error loss function.After adding the hinge-cross-entropy loss function, PSNR and SSIM improved by 1.61 dB and 1.57% in the DIV2 K dataset, respectively.In the six categories of the COCO dataset,PSNR increased by 0.51,0.64,0.33,0.69,0.56,0.24 dB,and SSIM increased by 2.24%,2.74%,2.04%,2.00%,3.7%,1.63%,respectively.This is because the partial derivative value of the MSE loss function will be very small when the output probability value is close to 0 or 1,which may cause the partial derivative value to almost disappear at the beginning of training of the model.As a result, the learning rate of the model is very slow at the beginning, and the use of cross-entropy as the loss function will not lead to such a situation.Therefore, compared to the effect without the hinged loss function and with the added the MSE loss function,the model using the hinge-cross-entropy loss function has a significant improvement in rendering quality.

    Table 2: Verification of the effectiveness of hinge-cross-entropy loss

    4.2 Quantitative Comparison

    Effect of improved self-attention moduleTo verify the effectiveness of the improved self-attention module, we added the self-attention module and the improved module to the pix2pix model and compared it with the original pix2pix model.The experimental results are shown in Fig.5, where SA’is the standard normal distribution in which the diagonal line conforms to the initial weight.The mean is 0 and the variance is 1 in the diagonal matrix,and SA”is the uniform random number in the diagonal matrix where the initial weight is 0–1.By comparing the results before and after adding hinge loss, we can see that compared with other models, the rendered image obtained by adding an improved self-attention module (pix2pix+SA’and SA’+Hinge+BCE) is closer to the original color image, with better rendering effect, clearer details and less rendering error.This is because the improved self-attention mechanism reduces the dependence on external information and uses the inherent information of features as much as possible for attentional interaction.In addition,the selfattention mechanism can effectively capture the feature dependence of long distance and extract the important information of global context.In comparison,our proposed approach SA’+Hinge+BCE has the best performance.Not only is the improved attention mechanism added,but the hinge-crossentropy loss function makes the gap between positive and negative samples large enough to produce an image that more closely resembles the real color image.

    Figure 5:Quantitative comparison

    The experimental results are shown in Table 3,and we can see that the rendering effect is better when the diagonal matrix with the initial weight conforming to the standard normal distribution with a mean of 0 and a variance of 1,namely SA’,is used.Compared with the original pix2pix model,the proposed method SA’+ Hinge+BCE improved 1.82% in the DIV2 K dataset, and 1.76 dB in terms of PSNR.In the six categories of COCO dataset, SSIM increased by 2.11%, 2.47%, 2.18%, 1.54%,3.73%,0.43%,and PSNR increased by 0.41,0.67,0.27,0.64,0.44,0.29 dB,respectively.Therefore,the improved self-attention module and the hinge-cross-entropy loss function can enhance the stability of the model and improve the rendering effect of the existing color rendering algorithm based on the GAN model.Therefore,the improved self-attention module and hinge-cross-entropy loss function in this paper are effective.It can enhance the stability of the model to different degrees and improve the existing image rendering algorithm based on the GAN model.

    Table 3: Verification of the effectiveness of the improved self-attention module

    Fig.6 shows the effect of our HCEGAN compared with the original pix2pix model on the COCO dataset.In order to improve the performance of the model,the current color rendering algorithm based on deep learning inevitably accumulates modules,which leads to the deep level and high complexity of the neural network.This paper realizes image color rendering based on GAN.Although the selfattention module is added to the model in this paper,the complexity of the system does not change much.At the same time, the addition of new loss function further strengthens the stability of the model.Therefore, the model’s lightweight and high performance are the limitations at present, and future work will also be carried out in this aspect.

    Figure 6:Our results compared with other models

    5 Conclusion

    At present,image color rendering based on the generative adversarial network has helped the medical industry to highlight and speed up the diagnosis of many diseases in the internet of medical things.Color images are beneficial because they can better highlight deep information in an image.In order to improve the existing GAN-based color model and render grayscale images,this paper introduces a new hinge-cross-entropy loss function and an improved self-attention module,and proposes a new hingecross-entropy GAN.In this paper,we use the DIV2 K and COCO datasets to verify the effectiveness of the proposed method and its superiority to prior approaches.The experimental results show that our hinge-cross-entropy GAN model demonstrates a great improvement in rendering quality and effect.Moreover,the stability of the model is greatly improved.However,the current GAN model has high model complexity and difficulty in pre-training.It is difficult to realize a model lightweight on the basis of ensuring algorithm efficiency.In the future,we will focus on achieving the high performance of GAN models while reducing model complexity.At the same time,we plan to extend the method to other tasks,such as style transfer and single image super-resolution.

    Acknowledgement:We extend our gratitude to the peer reviewers for their helpful comments on an earlier version of the paper.

    Funding Statement:The authors received National Natural Science Foundation of China (No.61902311) funding for this study.And the project was supported in part by the Natural Science Foundation of Shaanxi Province in China under Grants 2022JM-508,2022JM-317 and 2019JM-162.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    日日摸夜夜添夜夜爱| 六月丁香七月| 久久综合国产亚洲精品| 国产高清有码在线观看视频| 伦理电影免费视频| 欧美成人a在线观看| 最近的中文字幕免费完整| 国产久久久一区二区三区| 在线观看免费视频网站a站| 日韩,欧美,国产一区二区三区| 女性生殖器流出的白浆| 最近中文字幕2019免费版| 久久女婷五月综合色啪小说| 国产av码专区亚洲av| 欧美日韩综合久久久久久| 男女无遮挡免费网站观看| 高清av免费在线| 国产成人免费观看mmmm| 精品国产三级普通话版| 91久久精品电影网| 国产成人精品久久久久久| 老司机影院毛片| 亚洲第一区二区三区不卡| 日韩欧美一区视频在线观看 | 日韩人妻高清精品专区| 欧美区成人在线视频| 熟女电影av网| 1000部很黄的大片| 纵有疾风起免费观看全集完整版| 两个人的视频大全免费| 永久免费av网站大全| 亚洲精品日韩在线中文字幕| 欧美亚洲 丝袜 人妻 在线| 婷婷色麻豆天堂久久| 欧美日韩视频精品一区| 天美传媒精品一区二区| 欧美成人精品欧美一级黄| 日韩在线高清观看一区二区三区| 午夜福利网站1000一区二区三区| 九色成人免费人妻av| 成人无遮挡网站| 国产欧美日韩一区二区三区在线 | 中文在线观看免费www的网站| 黄色配什么色好看| 久久久久精品久久久久真实原创| 亚洲成人手机| 国产在线一区二区三区精| 免费看光身美女| 国产极品天堂在线| 久久人人爽av亚洲精品天堂 | 亚洲精华国产精华液的使用体验| 色吧在线观看| 观看av在线不卡| 我的女老师完整版在线观看| 国产亚洲最大av| 美女国产视频在线观看| 国国产精品蜜臀av免费| 偷拍熟女少妇极品色| 91精品伊人久久大香线蕉| 99久国产av精品国产电影| 一二三四中文在线观看免费高清| 国产av国产精品国产| 91在线精品国自产拍蜜月| 国产亚洲最大av| 美女xxoo啪啪120秒动态图| 久久久久久久精品精品| 久久 成人 亚洲| 一区二区三区乱码不卡18| 国内精品宾馆在线| 精品人妻熟女av久视频| 人妻系列 视频| 国产精品伦人一区二区| 国产深夜福利视频在线观看| 日韩电影二区| 国产精品.久久久| av播播在线观看一区| 搡老乐熟女国产| 99热6这里只有精品| 午夜免费男女啪啪视频观看| 免费人妻精品一区二区三区视频| 黄色怎么调成土黄色| 边亲边吃奶的免费视频| 日韩av在线免费看完整版不卡| 蜜臀久久99精品久久宅男| 精华霜和精华液先用哪个| 哪个播放器可以免费观看大片| 久久精品国产自在天天线| 亚洲怡红院男人天堂| 免费久久久久久久精品成人欧美视频 | 亚洲成人一二三区av| 亚洲国产精品国产精品| 蜜桃在线观看..| 夜夜看夜夜爽夜夜摸| 成人亚洲精品一区在线观看 | 美女视频免费永久观看网站| 中文字幕免费在线视频6| 日韩电影二区| 美女脱内裤让男人舔精品视频| 久久国产乱子免费精品| 国产精品欧美亚洲77777| 最后的刺客免费高清国语| 晚上一个人看的免费电影| 人人妻人人爽人人添夜夜欢视频 | 亚洲,欧美,日韩| 永久网站在线| 亚洲欧洲国产日韩| 啦啦啦中文免费视频观看日本| 狂野欧美白嫩少妇大欣赏| 夜夜爽夜夜爽视频| av福利片在线观看| 在线观看av片永久免费下载| 我的女老师完整版在线观看| 在线观看美女被高潮喷水网站| 简卡轻食公司| 国产免费又黄又爽又色| 人人妻人人爽人人添夜夜欢视频 | 女性被躁到高潮视频| 婷婷色av中文字幕| 高清在线视频一区二区三区| 在线播放无遮挡| 一级毛片黄色毛片免费观看视频| 少妇人妻一区二区三区视频| 精品久久国产蜜桃| 久久国产亚洲av麻豆专区| 狂野欧美白嫩少妇大欣赏| 欧美成人午夜免费资源| 国产av一区二区精品久久 | 国内揄拍国产精品人妻在线| 亚洲内射少妇av| 国产免费一区二区三区四区乱码| 干丝袜人妻中文字幕| 免费少妇av软件| 久久久久久久久久成人| 大陆偷拍与自拍| 国产黄频视频在线观看| 色网站视频免费| 欧美变态另类bdsm刘玥| 亚洲av二区三区四区| 91aial.com中文字幕在线观看| 国产一区亚洲一区在线观看| 人人妻人人添人人爽欧美一区卜 | 日日啪夜夜爽| 欧美变态另类bdsm刘玥| .国产精品久久| 2022亚洲国产成人精品| 日韩大片免费观看网站| 精品久久国产蜜桃| 久久青草综合色| 国产片特级美女逼逼视频| 欧美极品一区二区三区四区| 免费少妇av软件| 看免费成人av毛片| 这个男人来自地球电影免费观看 | 下体分泌物呈黄色| 蜜桃久久精品国产亚洲av| 日韩在线高清观看一区二区三区| 肉色欧美久久久久久久蜜桃| 亚洲精品,欧美精品| 十分钟在线观看高清视频www | 视频区图区小说| 我要看日韩黄色一级片| 18禁在线无遮挡免费观看视频| 午夜激情福利司机影院| 涩涩av久久男人的天堂| 欧美激情国产日韩精品一区| 欧美精品一区二区大全| 日韩不卡一区二区三区视频在线| 日韩强制内射视频| 久久国内精品自在自线图片| 蜜桃亚洲精品一区二区三区| 男人添女人高潮全过程视频| 黄色一级大片看看| 日本与韩国留学比较| 五月玫瑰六月丁香| 91精品伊人久久大香线蕉| 午夜激情福利司机影院| 一级爰片在线观看| 人妻制服诱惑在线中文字幕| 2022亚洲国产成人精品| 18禁裸乳无遮挡动漫免费视频| 久久综合国产亚洲精品| 国产 精品1| 国产69精品久久久久777片| 欧美极品一区二区三区四区| 日韩成人伦理影院| 制服丝袜香蕉在线| 国内揄拍国产精品人妻在线| 我要看日韩黄色一级片| 成年人午夜在线观看视频| 亚洲美女搞黄在线观看| 亚洲久久久国产精品| 天天躁夜夜躁狠狠久久av| 色婷婷av一区二区三区视频| 国产综合精华液| 亚洲国产精品成人久久小说| 精品亚洲乱码少妇综合久久| 99国产精品免费福利视频| 天天躁夜夜躁狠狠久久av| 97超视频在线观看视频| 亚洲成人手机| 如何舔出高潮| 熟女av电影| 在线看a的网站| 国产成人91sexporn| av卡一久久| 成人二区视频| 久久亚洲国产成人精品v| 欧美精品人与动牲交sv欧美| 99热网站在线观看| 亚洲精品久久午夜乱码| 国产69精品久久久久777片| 人妻 亚洲 视频| 在线天堂最新版资源| 99热国产这里只有精品6| 久久久久久久精品精品| 久久久久性生活片| 国产日韩欧美在线精品| av天堂中文字幕网| 制服丝袜香蕉在线| 天堂俺去俺来也www色官网| 直男gayav资源| 日韩,欧美,国产一区二区三区| 永久网站在线| 哪个播放器可以免费观看大片| 亚洲精品日韩在线中文字幕| 成人毛片60女人毛片免费| 亚洲一级一片aⅴ在线观看| 欧美xxxx黑人xx丫x性爽| 久久久久久人妻| 伊人久久精品亚洲午夜| 五月伊人婷婷丁香| 校园人妻丝袜中文字幕| 欧美丝袜亚洲另类| 成人一区二区视频在线观看| 午夜福利在线观看免费完整高清在| 99热网站在线观看| 欧美日韩国产mv在线观看视频 | 中文字幕av成人在线电影| 各种免费的搞黄视频| 美女福利国产在线 | 国产精品一区二区在线观看99| 99热国产这里只有精品6| 久久 成人 亚洲| 女的被弄到高潮叫床怎么办| 99国产精品免费福利视频| 欧美日韩在线观看h| 舔av片在线| 久久精品久久久久久久性| 亚洲欧美成人精品一区二区| 亚洲av中文av极速乱| 伦理电影免费视频| 久久精品久久久久久久性| 高清视频免费观看一区二区| 亚洲精品日韩在线中文字幕| 97在线视频观看| 97精品久久久久久久久久精品| 国产精品人妻久久久久久| 亚洲精品久久久久久婷婷小说| 久久ye,这里只有精品| 51国产日韩欧美| 国产又色又爽无遮挡免| 亚洲美女黄色视频免费看| 久久人人爽av亚洲精品天堂 | 国产真实伦视频高清在线观看| 国产精品女同一区二区软件| 美女视频免费永久观看网站| 国产一区二区三区综合在线观看 | 久久6这里有精品| 国产精品久久久久久久久免| 人妻制服诱惑在线中文字幕| 久久久久精品久久久久真实原创| 成人无遮挡网站| 国产色爽女视频免费观看| 精品国产露脸久久av麻豆| 一本—道久久a久久精品蜜桃钙片| 精品久久久精品久久久| 极品少妇高潮喷水抽搐| av在线蜜桃| 少妇的逼好多水| 在线免费观看不下载黄p国产| 免费看不卡的av| av在线观看视频网站免费| 欧美日韩视频高清一区二区三区二| 性色av一级| 免费看光身美女| 精品国产三级普通话版| 午夜老司机福利剧场| 观看免费一级毛片| 天堂中文最新版在线下载| 欧美另类一区| 热re99久久精品国产66热6| 99re6热这里在线精品视频| av国产免费在线观看| 91精品伊人久久大香线蕉| 亚洲国产精品国产精品| 1000部很黄的大片| 国产亚洲最大av| 国产精品伦人一区二区| 亚洲色图av天堂| 日本vs欧美在线观看视频 | 少妇 在线观看| 欧美+日韩+精品| 少妇人妻 视频| 春色校园在线视频观看| 成年女人在线观看亚洲视频| 91久久精品电影网| 欧美+日韩+精品| 日韩中字成人| 国产精品久久久久久av不卡| 国产一区有黄有色的免费视频| 欧美三级亚洲精品| 国产精品欧美亚洲77777| 韩国高清视频一区二区三区| 久久鲁丝午夜福利片| 国产成人91sexporn| 国产精品一区www在线观看| 中文资源天堂在线| 一区在线观看完整版| 欧美成人一区二区免费高清观看| 少妇精品久久久久久久| 91午夜精品亚洲一区二区三区| 黄色怎么调成土黄色| 亚洲图色成人| 我的老师免费观看完整版| 欧美少妇被猛烈插入视频| 男人狂女人下面高潮的视频| 亚洲经典国产精华液单| 国产成人精品婷婷| 麻豆乱淫一区二区| 岛国毛片在线播放| 亚州av有码| 亚洲精品成人av观看孕妇| 国产一区二区三区综合在线观看 | 日韩免费高清中文字幕av| 夜夜爽夜夜爽视频| 亚洲精品,欧美精品| 成人无遮挡网站| 美女中出高潮动态图| 欧美最新免费一区二区三区| 久久久欧美国产精品| 一区二区三区精品91| 大片免费播放器 马上看| 国产深夜福利视频在线观看| 蜜桃在线观看..| 久久99精品国语久久久| 国产伦理片在线播放av一区| 成人午夜精彩视频在线观看| 日产精品乱码卡一卡2卡三| 亚洲精品色激情综合| 春色校园在线视频观看| 99久久精品一区二区三区| 久久国产乱子免费精品| 插阴视频在线观看视频| 国产欧美日韩精品一区二区| 91aial.com中文字幕在线观看| 18禁在线无遮挡免费观看视频| av黄色大香蕉| 边亲边吃奶的免费视频| 国产精品欧美亚洲77777| 热re99久久精品国产66热6| 尤物成人国产欧美一区二区三区| 成人二区视频| 久久久久久久久久人人人人人人| 免费久久久久久久精品成人欧美视频 | 久久久久精品久久久久真实原创| 免费观看的影片在线观看| 少妇的逼水好多| 精品国产乱码久久久久久小说| av不卡在线播放| 亚洲精华国产精华液的使用体验| 精品国产三级普通话版| 亚洲国产色片| 国产精品99久久久久久久久| 一本色道久久久久久精品综合| 亚洲精品成人av观看孕妇| 三级经典国产精品| 啦啦啦在线观看免费高清www| 美女高潮的动态| 天堂中文最新版在线下载| 18禁在线播放成人免费| 97在线人人人人妻| 国产高潮美女av| 天美传媒精品一区二区| 少妇精品久久久久久久| 欧美日韩在线观看h| 亚洲欧美清纯卡通| 色5月婷婷丁香| 九九久久精品国产亚洲av麻豆| 日韩 亚洲 欧美在线| 一个人看的www免费观看视频| 少妇的逼好多水| 中文字幕制服av| 寂寞人妻少妇视频99o| 亚洲,一卡二卡三卡| 亚洲av成人精品一区久久| 只有这里有精品99| 最近手机中文字幕大全| 亚洲无线观看免费| 成人国产av品久久久| 丰满迷人的少妇在线观看| 新久久久久国产一级毛片| 国产乱人视频| 国产成人免费无遮挡视频| 久久这里有精品视频免费| 精品一区二区免费观看| 99热6这里只有精品| 国产av精品麻豆| 狂野欧美激情性xxxx在线观看| 久久精品国产a三级三级三级| 汤姆久久久久久久影院中文字幕| 欧美成人一区二区免费高清观看| 亚洲精华国产精华液的使用体验| 亚洲av.av天堂| 色视频www国产| 永久免费av网站大全| 在线观看免费高清a一片| 亚洲人成网站在线观看播放| 精品一区二区免费观看| 久久久久性生活片| 男女免费视频国产| 国精品久久久久久国模美| av专区在线播放| 久久99热这里只有精品18| 美女cb高潮喷水在线观看| 不卡视频在线观看欧美| 午夜激情久久久久久久| 熟妇人妻不卡中文字幕| 午夜日本视频在线| 亚洲精品国产av成人精品| 亚洲性久久影院| 国产男女内射视频| 特大巨黑吊av在线直播| 欧美精品人与动牲交sv欧美| 大片免费播放器 马上看| 中文资源天堂在线| 自拍欧美九色日韩亚洲蝌蚪91 | 老女人水多毛片| 黄片wwwwww| 97精品久久久久久久久久精品| 欧美最新免费一区二区三区| 成人18禁高潮啪啪吃奶动态图 | 亚洲精品视频女| 91精品伊人久久大香线蕉| 国产成人午夜福利电影在线观看| 亚洲av男天堂| 亚洲国产精品专区欧美| 精品一区二区三区视频在线| 五月开心婷婷网| 免费大片18禁| 一级黄片播放器| 色视频www国产| 王馨瑶露胸无遮挡在线观看| 啦啦啦中文免费视频观看日本| 黄色欧美视频在线观看| 男女下面进入的视频免费午夜| a级毛色黄片| 亚洲av欧美aⅴ国产| 男男h啪啪无遮挡| 亚洲性久久影院| 高清av免费在线| 高清在线视频一区二区三区| 中文字幕av成人在线电影| 18禁裸乳无遮挡免费网站照片| 久久精品国产亚洲av涩爱| 2018国产大陆天天弄谢| 亚洲人成网站在线播| 亚洲真实伦在线观看| 黄色视频在线播放观看不卡| 精品国产一区二区三区久久久樱花 | 免费观看a级毛片全部| 水蜜桃什么品种好| 欧美三级亚洲精品| 欧美成人一区二区免费高清观看| 亚洲久久久国产精品| 亚洲电影在线观看av| 国产精品无大码| 夜夜骑夜夜射夜夜干| 久久精品国产亚洲av涩爱| 久久久精品94久久精品| 高清在线视频一区二区三区| 深夜a级毛片| 国产免费视频播放在线视频| 一区二区三区四区激情视频| 亚洲丝袜综合中文字幕| 日韩精品有码人妻一区| 韩国高清视频一区二区三区| 亚洲精品aⅴ在线观看| 男男h啪啪无遮挡| 久久av网站| 欧美另类一区| 一级爰片在线观看| 国产伦精品一区二区三区视频9| 在线免费观看不下载黄p国产| 亚洲成色77777| 国产综合精华液| 日韩欧美一区视频在线观看 | 黄色视频在线播放观看不卡| 黄色一级大片看看| 舔av片在线| 精品亚洲乱码少妇综合久久| 一级爰片在线观看| 欧美bdsm另类| 2021少妇久久久久久久久久久| 男人爽女人下面视频在线观看| 久久鲁丝午夜福利片| 一级毛片电影观看| 赤兔流量卡办理| 亚洲精品国产成人久久av| 国产又色又爽无遮挡免| 成人18禁高潮啪啪吃奶动态图 | 国产精品一区二区性色av| 综合色丁香网| 一个人看视频在线观看www免费| 久久综合国产亚洲精品| 亚洲人成网站在线观看播放| 亚洲内射少妇av| 国产成人午夜福利电影在线观看| 精品久久国产蜜桃| 久久婷婷青草| 自拍欧美九色日韩亚洲蝌蚪91 | 国产乱人偷精品视频| 国产极品天堂在线| 99热6这里只有精品| 亚洲精品久久久久久婷婷小说| 久久久久国产精品人妻一区二区| 久久女婷五月综合色啪小说| 午夜免费鲁丝| 亚洲国产精品999| 我的老师免费观看完整版| 亚洲精品色激情综合| 中文字幕制服av| 99热国产这里只有精品6| 午夜精品国产一区二区电影| 黄片无遮挡物在线观看| 日韩 亚洲 欧美在线| 久久精品人妻少妇| 水蜜桃什么品种好| 国产精品一区二区性色av| 最后的刺客免费高清国语| 一个人看的www免费观看视频| 有码 亚洲区| 免费av中文字幕在线| 国产乱人偷精品视频| 最黄视频免费看| 国产欧美日韩一区二区三区在线 | 妹子高潮喷水视频| 天堂中文最新版在线下载| 亚洲激情五月婷婷啪啪| 国产精品久久久久久精品电影小说 | 身体一侧抽搐| 热99国产精品久久久久久7| 亚洲电影在线观看av| 亚洲综合精品二区| 黑丝袜美女国产一区| 国产免费一区二区三区四区乱码| 老司机影院毛片| 日本一二三区视频观看| 午夜福利在线观看免费完整高清在| 国产亚洲精品久久久com| 欧美xxⅹ黑人| 中文乱码字字幕精品一区二区三区| 男的添女的下面高潮视频| 人妻少妇偷人精品九色| 交换朋友夫妻互换小说| 狂野欧美激情性bbbbbb| 国产精品偷伦视频观看了| 热re99久久精品国产66热6| 少妇人妻一区二区三区视频| 蜜桃亚洲精品一区二区三区| 免费播放大片免费观看视频在线观看| 高清毛片免费看| 久久久成人免费电影| av黄色大香蕉| 爱豆传媒免费全集在线观看| 久久久久久久亚洲中文字幕| 久久精品国产自在天天线| 超碰97精品在线观看| 成人黄色视频免费在线看| 国产精品蜜桃在线观看| 九九久久精品国产亚洲av麻豆| 国产亚洲av片在线观看秒播厂| 美女国产视频在线观看| 亚洲成人一二三区av| 日日啪夜夜爽| 亚洲天堂av无毛| 亚洲不卡免费看| 国内揄拍国产精品人妻在线| 亚洲精品一区蜜桃| 成人国产麻豆网| 日日啪夜夜爽| 国产男人的电影天堂91| 国产 一区 欧美 日韩| 亚洲av男天堂| 观看免费一级毛片| 免费播放大片免费观看视频在线观看| 久久精品久久久久久久性| 日韩伦理黄色片| 久久精品国产鲁丝片午夜精品| 国产精品av视频在线免费观看| 丝瓜视频免费看黄片| 精品人妻偷拍中文字幕| 丝袜脚勾引网站| 男人舔奶头视频| 人妻系列 视频| 久久99蜜桃精品久久| 女的被弄到高潮叫床怎么办| 亚洲综合色惰| 乱码一卡2卡4卡精品| 国产精品国产三级国产av玫瑰| 国产欧美日韩一区二区三区在线 | 国产精品一区二区在线不卡| 亚洲国产毛片av蜜桃av| 亚洲精品色激情综合| 国产精品麻豆人妻色哟哟久久| 99久久中文字幕三级久久日本|