• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Deep3DSketch-im: rapid high-fidelity AI 3D model generation by single freehand sketches*

    2024-03-06 09:17:10TianrunCHENRunlongCAOZejianLIYingZANGLingyunSUN

    Tianrun CHEN, Runlong CAO,Zejian LI,Ying ZANG,Lingyun SUN

    1College of Computer Science and Technology, Zhejiang University, Hangzhou 310027, China

    2School of Software Technology, Zhejiang University, Hangzhou 310027, China

    3School of Information Engineering, Huzhou University, Huzhou 313000, China

    E-mail: tianrun.chen@zju.edu.cn; crl1567@163.com; zejianlee@zju.edu.cn; 02750@zjhu.edu.cn; sunly@zju.edu.cn

    Received Apr.30, 2023; Revision accepted Nov.26, 2023; Crosschecked Jan.15, 2024

    Abstract: The rise of artificial intelligence generated content (AIGC) has been remarkable in the language and image fields, but artificial intelligence (AI) generated three-dimensional (3D) models are still under-explored due to their complex nature and lack of training data.The conventional approach of creating 3D content through computer-aided design (CAD) is labor-intensive and requires expertise, making it challenging for novice users.To address this issue,we propose a sketch-based 3D modeling approach,Deep3DSketch-im,which uses a single freehand sketch for modeling.This is a challenging task due to the sparsity and ambiguity.Deep3DSketch-im uses a novel data representation called the signed distance field (SDF) to improve the sketch-to-3D model process by incorporating an implicit continuous field instead of voxel or points, and a specially designed neural network that can capture point and local features.Extensive experiments are conducted to demonstrate the effectiveness of the approach, achieving state-of-the-art(SOTA)performance on both synthetic and real datasets.Additionally,users show more satisfaction with results generated by Deep3DSketch-im, as reported in a user study.We believe that Deep3DSketch-im has the potential to revolutionize the process of 3D modeling by providing an intuitive and easy-to-use solution for novice users.

    Key words: Content creation; Sketch; Three-dimensional (3D) modeling; 3D reconstruction; Shape from X;Artificial intelligence (AI)

    1 Introduction

    The abilities of artificial intelligence (AI) content generation have made significant strides in recent years, with notable progress made in generating images and text (Zhou et al., 2023; Huang and Wang, 2024; Lei and Li, 2024).However, despite the increasing demand for three-dimensional (3D)models in various applications, most existing artificial intelligence generated content (AIGC) methods have been focused on language and two-dimensional(2D)image generation,resulting in a significant lack of progress in generating 3D content.Generating 3D content poses several significant challenges, such as the need for 3D spatial awareness and the difficulty in representing complex 3D shapes.Furthermore, the lack of readily available 3D content data for training models is a significant challenge in the development of AI-generated 3D content-for years, researchers have to use computer-aided design (CAD)based approaches to create 3D models.However,the complexity and steep learning curve of CAD software pose a significant challenge for novice users, as observed in Chester (2007).The time-consuming and labor-intensive nature of the CAD-based approach also limits its scalability and hinders its potential to democratize 3D modeling, as noted by Reddy and Rangadu (2018).Recently, software like Tinkercad offers better usability, but the strategic knowledge(decomposing steps of constructing models)is still a challenge for users to learn(Mahapatra et al.,2019).As a result, there is a growing need for more intuitive and user-friendly 3D modeling methods that can better serve the needs of a broader range of users.

    To address the challenges of 3D content generation and the limitations of CAD-based approaches,this study explores sketch-based 3D modeling as a promising alternative.By leveraging the intuitive and natural form of computer-human interaction,we aim to allow users to create 3D models using freehand sketches as input, thereby greatly simplifying the learning progress of 3D modeling, thus resulting in an increase in the amount of quality 3D content produced.

    Existing sketch-based 3D modeling approaches are far from perfect.Many existing approaches require precise line drawings from multiple perspectives or follow a step-by-step workflow that assumes a strategic understanding of 3D modeling (Cohen et al., 1999; Deng et al., 2020).These methods, although effective, can be challenging for novice users and time-consuming.Additionally,other approaches that employ template primitives or retrieval-based techniques (Chen DY et al., 2003; Wang F et al.,2015) lack the customizability which is necessary to allow users to fully express their creative ideas.Thus,a more balanced approach is needed to provide both ease of use and flexibility for novice users, thereby enabling them to create custom 3D models with the minimum effort and maximum creativity.

    To accomplish the task of efficient and userfriendly 3D modeling, our approach aims to generate a detailed 3D model from a single freehand sketch input.This is a challenging task as the input is limited to a single sketch with the minimum information.Previous studies have used deep neural networks to achieve the sketch-to-model translation, using an encoder-decoder architecture that compresses the input sketch into a coarse representation (latent code) capturing information such as the semantic category and conceptual shape,followed by recovery of the 3D shapes using a decoder that calculates the offsets of a given number of points or vertices(Guillard et al.,2021;Zhang SH et al.,2021;Chen TR et al., 2023a, 2023b, 2023c; Zang et al.,2023).However, these approaches have difficulties in capturing intricate details due to the significant domain gap between a sketch and a 3D shape domain, and the limited resolution in points or vertex representations.

    Here, as shown in Fig.1, we propose Deep3-DSketch-im, which can elevate the resolution of sketch-to-3D modeling to a new level.This resolution enhancement is enabled by different data representations-those other than the aforementioned points or voxels.Securing an integration of the implicit 3D surface representation, namely signed distance functions, into the sketch-to-model process would be the first step toward achieving this enhancement.The signed distance function encodes the distance of each point sample in 3D from the boundary of the shape, with a sign indicating whether the point is inside or outside the shape(Fig.2).For our sketch-to-model task, a convolutional neural network(CNN)first encodes the input sketch into a feature vector.Then, we use this feature vector to predict the signed distance field(SDF)value of a 3D point.By sampling different points with infinite possible locations, Deep3DSketch-im generates an implicit field of the underlying surface with infinite resolution.

    We have performed extensive experiments using synthetic and real hand-drawn datasets.Experimental results show that our approach achieves stateof-the-art (SOTA) performance.Our approach can reconstruct more details with higher-fidelity results.Our user study also shows that users are more satisfied with the models obtained by our approach.We believe that our approach allows for a wide range of applications, such as 3D printing, virtual reality, and video game development.Additionally, the ability to create custom 3D models quickly and easily could have significant impacts on industries such as architecture, product design, and entertainment.Deep3DSketch-im could become a powerful tool for democratizing 3D modeling, making it accessible to a wider range of users and potentially revolutionizing the way by which we design and create in 3D.

    Fig.1 Pipeline of our sketch-based 3D modeling approach

    Fig.2 Illustration of the signed distance field (SDF):(a) rendered 3D surface with S =0; (b) cross-section of the SDF

    2 Related works

    2.1 Sketch-based 3D modeling

    Sketch-based 3D modeling has been an active research area for many years, with numerous approaches proposed by researchers.One category of sketch-based 3D modeling approaches is interactive approaches, which requires breaking down the task into sequential steps or using specific drawing gestures or annotations.These methods have been shown to require significant strategic knowledge, making them challenging for novice users.For instance, Li CJ et al.(2020) used a two-stage approach for coarse-to-fine reconstruction, while Cohen et al.(1999) used annotation-based feedback to refine the 3D model.

    In contrast,end-to-end approaches such as template primitives or retrieval-based methods tend to be more straightforward but lack customizability.These approaches involve generating the 3D model directly from the sketch without any intermediate steps.For example, Chen DY et al.(2003) used 3D geometric primitives for sketch-based modeling,while Wang F et al.(2015) introduced a retrievalbased approach that uses a database of 3D models to find the closest match to the input sketch.Recently,deep learning based approaches have been proposed for single-view 3D reconstruction, including sketchbased 3D modeling.For example, Zhang SH et al.(2021) and Chen TR et al.(2023a, 2023b, 2023c)proposed the use of an encoder-decoder backbone to output the offset of a round shape template,but the adoption of such an approach is able to result in only the representation of coarse shapes characterized by a lack of structural details.Gao et al.(2022) used density maps and point clouds as the representation;though the representation is capable of generating high-fidelity results, further processing needs to be done for most applications requiring mesh representation.Guillard et al.(2021)proposed methods that reconstruct the 3D model with a two-stage refinement scheme,but it cannot provide high-fidelity generation in real time.Moreover,these approaches face substantial challenges due to the sparse and abstract nature of sketches lacking fine boundary information and texture information for depth estimation, making it difficult to produce high-quality 3D shapes.In Zhong et al.(2020),these challenges were illustrated and analyzed in detail.

    In this study, we introduce a novel approach called Deep3DSketch-im, which leverages SDF to represent 3D surfaces,rather than using point clouds or voxels, to generate higher-fidelity 3D models.By incorporating SDFs,our approach can capture more structural details in the input sketches and produce models with a theoretically infinite resolution.This sets our approach apart from existing methods, as it does not require a fixed topology assumption and can produce accurate ground truths without approximating metrics(Xu et al.,2022,2023;Lin GY et al.,2023; Yang et al., 2023).This makes it suitable for both novice and experienced users who want to create high-quality 3D models from sketches.

    2.2 Single-view 3D reconstruction

    The task of reconstructing 3D geometry from a single 2D image has been a challenging problem in the fields of computer vision and computer graphics for many years.In recent years, data-driven approaches have gained popularity with the advent of large-scale datasets like ShapeNet (Chang et al.,2015).Some works(Chen ZQ and Zhang,2019;Park et al., 2019) use category-level information to infer 3D representations, while others (Kato et al., 2018;Liu et al.,2019a,2019b)directly generate 3D models from 2D images using differentiable rendering techniques.More recently, unsupervised methods for implicit function representations using differentiable rendering have been proposed (Lin CH et al., 2020;Yu et al.,2021).

    However,most of these methods focus on learning 3D geometry from 2D colored images.In contrast,our approach aims to generate 3D meshes from 2D sketches, which are a more abstract and sparse form of image representation.Sketches lack important information like texture, lighting, and shading,making it challenging to infer 3D geometry accurately (Chen TR et al., 2023a, 2023b; Zang et al.,2023; Zhang SZ et al., 2023).Additionally, sketches are often incomplete, and the same set of strokes can have different interpretations in 3D,adding ambiguity to the problem.Therefore, it is critical to develop a method that can accurately interpret and reconstruct 3D shapes from sparse and ambiguous sketches.

    In this study,we propose a novel approach that addresses these challenges and provides an efficient and accurate solution for sketch-based 3D modeling.Our approach uses a deep learning based approach that learns to interpret the abstract representation of sketches and reconstructs a high-quality 3D mesh.We show that our approach outperforms SOTA methods on benchmark datasets for sketchbased 3D modeling.

    3 Method

    3.1 Signed distance field

    Our objective is to generate a highly detailed and accurate 3D model of an object from a given image.To achieve this,we adopt a novel approach that represents the 3D shape as an SDF (Tong X, 2022).By representing the shape as an SDF,we can model the object’s surface as a level set of the function, allowing us to generate a high-resolution mesh of the object’s surface.As illustrated in Fig.2, the signed distance function is a continuous function that maps a given spatial pointp= (x,y,z)∈R3to a real value:s=SDF(p), whereSDFrefers to the signed distance function.In contrast to commonly used 3D representations such as depth, the absolute value ofSDF(p) indicates the distance of a pointpto the surface, and the sign ofSDF(p) indicates whetherpis inside or outside the surface.The iso-surfaceS′={p|SDF(p)=0}implicitly represents the 3D shape.In implementation,we first define a dense 3D grid and predict signed distance function values for each point in the grid.With these values calculated,we can then use the Marching Cubes algorithm to obtain the 3D mesh that corresponds to the iso-surfaceS′.

    3.2 Network architecture

    As illustrated in Fig.3a, the conventional sketch-to-3D modeling approach is composed of an encoder-decoder network structure, where the encoderEtransforms the sparse and ambiguous input sketch into a latent shape codezsthat summarizes the sketch at a coarse level.The decoderDis then used to transfer the latent shape codezsto the meshMΘ=D(zs).Nevertheless,this design has the limitation that only coarse shapes are generated when, owing to the fact of the number of vertices or points being limited,the resolution is low.Moreover,the significant domain gaps between sketches and 3D shapes make it challenging to generate high-fidelity 3D shapes.

    On the contrary, our approach involves projecting each 3D query point onto the image plane and gathering multiscale CNN features for the corresponding image patch.The collected features are then used by the neural network to decode the given spatial point into an SDF value,using the multiscale local image features.

    Fig.3 Network architecture: (a) existing encoder-decoder structure of neural networks; (b) pipeline of our Deep3DSketch-im, which samples points at mesh and projects the points onto the image plane

    The details of our network structure are illustrated in Fig.3b.The input sketch can be seen as a binary imageI ∈{0,1}W×H(WandHare the width and height of the sketch, respectively), in whichI[i,j]=0 if marked by the stroke, andI[i,j]=1 otherwise.The image of the sketch is fed into an image encoder for multiscale image feature extraction,obtaining feature maps at different scales.Meanwhile, we project a 3D pointp ∈R3onto the image plane, obtaining a 2D locationq ∈R2.We obtain the local image features by retrieving the features on each feature map corresponding to locationqand concatenating them.As the feature maps in the later layers have smaller dimensions than the original image, we resize them to the original size with bilinear interpolation and extract the resized features at locationq.

    Additionally, as proposed in Wang WY et al.(2019),we extract point features for each pointqby applying 1×1 convolution and rectified linear unit(ReLU) activation functions after the convolution,resulting in feature vectors of increased length.We concatenate the point features and local features into a new feature vector, and then feed the new feature vector into a decoder that involves multiple 1×1 convolutions and ReLU activation functions to obtain the SDF prediction value.The local feature is used to capture detailed information in the sketch input, which is demonstrated effectively in our later experiment.

    3.3 Sketch view prediction

    Note that the abovementioned network needs the sketch pose as the input.However, for a more user-friendly experience, it is better to allow users to use the network without additional input.We find that we can predict pose based on synthesized data.Specifically, following the approach adopted in Chen TR et al.(2023a, 2023c), we design a separate pose-estimation network and train it in a fully supervised manner.We use an encoderEto produce latent codezlfrom the sketches and input it to the viewpoint prediction module, which consists of two fully connected layersDvaimed at producing the viewpoint estimateξpred, represented by an Euler angle.The viewpoint prediction module is optimized in a fully supervised manner with the input of the ground-truth (GT) viewpointξgt, supervised by a viewpoint prediction lossLv, which adopts the mean-squared error(MSE)loss for the predicted and GT poses,defined as

    The viewpoint prediction model is trained along with a 3D model generation process as in Chen TR et al.(2023b, 2023c).In the experiment, we find that the predicted pose can effectively guide the network with very little performance drop.

    3.4 Loss function

    We adopt a continuous signed distance function regression approach that allows us to extract surfaces corresponding to different iso-values as in Wang WY et al.(2019).To focus the network on recovering details near and inside the iso-surface, we use a weighted loss function.The loss function is defined as

    4 Experiments

    4.1 Dataset

    There is a limited availability of datasets that include both sketches and their corresponding 3D models for research purposes.Zhang SH et al.(2021), recently used synthetic data from the ShapeNet-Synthetic dataset for their training data.The ShapeNet-Synthetic dataset includes 13 categories of 3D objects, and synthetic data are generated using a canny edge detector on rendered images from Kar et al.(2017).The trained model is then evaluated on real-world data from the ShapeNet-Sketch dataset, which includes 1300 sketches and their corresponding 3D shapes.These sketches are drawn by human volunteers with varying levels of skills,based on images of 3D objects from Kar et al.(2017).

    4.2 Implementation details

    The image encoder of the sketch is VGG-16.The decoder consists of three layers,with 1×1 convolution followed by an ReLU function on the first two layers, and no activation function on the last layer,which is a 1×1 value prediction.The decoder takes the concatenation of point features and local multiscale features.During training, we focus on points near the iso-surfaceS′,which is achieved by employing Monte Carlo sampling.Specifically,we randomly select 2048 grid points from a Gaussian distributionN(0,0.1).Additionally,we set the parametersm1=4,m2=1,andδ=0.01 in Eq.(2)to ensure effective recovery of details near and inside the iso-surface.Our network is implemented using PyTorch and optimized using the Adam optimizer with a learning rate of 1×10-4and a batch size of 16, and employs an NVIDIA GeForce RTX 3090 Graphics Card.

    4.3 Experimental results and performance comparison

    We assess the performance of our method by comparing it with that of the SOTA model, following the same protocol as in Zhang SH et al.(2021).The model is trained for each category.We use the official training-evaluation-testing split and evaluate both the ShapeNet-Synthetic (edge-detected sketch)and ShapeNet-Sketch(hand-drawn sketches)datasets.In Table 1,the ShapeNet-Synthetic dataset is used for experiments,which provides accurate GT 3D models for training and evaluation.To evaluate the fidelity of the generated meshes, we employ the Chamfer distance metric, which is a widely used measure for 3D reconstruction.Some latest baseline approaches, namely Sketch2Model (Zhang SH et al., 2021), Deep3DSketch (Chen TR et al.,2023a),Sketch2Mesh(Guillard et al.,2021),and the deep implicit surface network (DISN) (Wang WY et al.,2019)are used for comparison.For a fair comparison, we use only the first feed-forward stage of Sketch2Mesh and do not perform the post-processing optimization step (shown as Sketch2Mesh).In the experiment, our approach demonstrates high effectiveness and achieves SOTA performance.The visualization presented in Fig.4 also emphasizes the significant performance elevation of our proposed Deep3DSketch-im.

    In Table 2,we further evaluate the performance in relation to real-world human drawings through the ShapeNet-Sketch dataset.Due to the limited number of samples, we train the model on the ShapeNet-Synthetic dataset and use the ShapeNet-Sketch dataset for evaluation.The results also show that our approach can consistently produce higherfidelity results when it comes to real hand-drawn datasets, as illustrated in Fig.5.Particularly, many detailed structures are accurately captured by our proposed Deep3DSketch-im, for example, the top and the side mirror of the car.The results show that Deep3DSketch-im is a robust network that can generalize well in real-world data, while it is trained only on synthetic data from edge detectors.The visualization results illustrate the obvious effectiveness of Deep3DSketch-im in producing models with higher quality and fidelity in structure.Further research can be conducted in terms of domain generalization and domain adaptation to better mitigate the domain gap between the synthetic and real-world data(Tong YZ et al.,2023;Zhu et al.,2023a,2023b,2023c)or collect more real data for training.

    4.4 Using Deep3DSketch-im w/o view input

    As mentioned above,despite inputting GT pose during the evaluation in Tables 1 and 2, we argue that the view information is not needed as the userinput because the view prediction network can predict the view and input the same to the model generation process, which is also beneficial for user experience in real-world applications.We test the performance with no GT pose but only predicted poses.As shown in Fig.6,Deep3DSketch-im without viewpoint as the input works as well as the network with GT poses.The quantitative evaluation results are shown in Table 3.

    Table 1 Quantitative evaluation of the ShapeNet-Synthetic dataset

    Table 2 Quantitative evaluation of the ShapeNet-Sketch dataset

    4.5 Runtime&time complexity evaluation

    Once the network is adequately trained,we measure its performance on a computer equipped with a consumer graphics card (NVIDIA GeForce RTX 3090).Our approach demonstrates a generation speed of 0.97 s,which is sufficient for fluent humancomputer interaction.

    Fig.4 Qualitative evaluation with state-of-the-art (SOTA) method in a synthesized dataset

    Fig.5 Qualitative evaluation with state-of-the-art (SOTA) method for a real-world dataset

    Note that in contrast with the approaches adopted in some NeRF studies (Fu et al., 2022;Metzer et al., 2022; Jo et al., 2023), our approach is a generalized approach that does not require timeconsuming per-object optimization.The time complexity is related to the number of points.If we assume that the voxelized grid used to represent the target model has a size ofN×N×N,whereNrespresents the resolution or density of the grid in each dimension, our algorithm’s complexity isO(N3) as we calculate the SDF value for each grid cell.

    4.6 User study

    To further assess the effectiveness of our sketchto-model algorithm,we carry out a user study using the mean option score (MOS) metric, which ranges from 1 to 5 (Seufert, 2019).The experiments follow the settings adopted in various studies (Cai et al.,2021; Michel et al., 2022; Yao et al., 2022; Chen TR et al., 2023c).We present 36 3D models generated by our algorithm to 12 designers who are well-versed in 3D content.The designers are asked to rate themodels based on two factors:

    Table 3 Quantitative evaluation of ShapeNet-Synthetic dataset w/o view input

    Fig.6 Qualitative evaluation with predicted pose

    Q1: How well does the output 3D model match the input sketch(fidelity)?

    Q2: How do you rate the quality of the output 3D model (quality)?

    Before the experiment,each participant is given a brief one-on-one introduction to the concepts of fidelity and quality.We average the scores and report the rating results in Table 4.The results indicate that our method outperforms the SOTA method in terms of users’subjective ratings.

    5 Real-world implication of the proposed approach

    Sketching is the most natural and intuitive way to express new ideas,and a single-view sketch is definitely the easiest approach for users to implement their thoughts.Although the downstream application is not the particular focus of the study,we noteone particular application of sketch-based 3D modeling that is rapid home interior design,which is the follow-up work to Chen TR et al.(2023c).With the involvement of context information,users can design the 3D model using sketches and place models within a real environment for a more immersive experience.In other words, the sketch can precisely define the six-dimensional(6D)pose and position of the generated object.We have demonstrated that the sketchguided approach is more efficient and easier to use compared to a “touch-based” approach concerning users’utility in terms of manipulating the generated 3D objects within the scene.Note that future research can focus on expanding the application of this sketch-based modeling tool, with appropriate shape representations (Xu et al., 2022; Yang et al., 2023;Zang et al., 2023), to eventually CAD/computeraided manufacturing(CAM)and many other fields.

    Table 4 Mean opinion scores (1-5) for Q1 (fidelity)and Q2 (quality)

    6 Conclusions

    We have presented a novel deep learning network, Deep3DSketch-im, for generating 3D models from a single 2D sketch.Considering the lack of fine details characterizing the existing sketch-tomodel approaches, we first introduce SDF to represent the 3D shape for infinite resolution.We design a network to capture the local features for fine-grained structure 3D modeling.Through experiments on the ShapeNet-Synthetic dataset,we have shown that our approach outperforms SOTA methods in terms of both quantitative metrics and user study ratings.Our method also expands the existing language- or vision-centered AIGC tools.We believe that our study opens up exciting possibility for creating 3D content from simple 2D sketches, which can have significant applications in industries such as gaming,animation,and architecture.

    Contributors

    Tianrun CHEN designed the research.Tianrun CHEN and Runlong CAO processed the data and performed the experiments.Tianrun CHEN drafted the paper.Zejian LI,Ying ZANG, and Lingyun SUN revised and finalized the paper.

    Compliance with ethics guidelines

    Lingyun SUN is an editor-in-chief assistant of this special issue, and he was not involved with the peer review process of this paper.All the authors declare that they have no conflict of interest.

    Data availability

    Our project data can be found at https://tianrunchen.github.io/Deep3DSketch-im.Other data that support the findings of this study are available from the corresponding authors upon reasonable request.

    卡戴珊不雅视频在线播放| 亚洲精品一二三| 欧美日韩一区二区视频在线观看视频在线| 免费播放大片免费观看视频在线观看| 人人妻人人添人人爽欧美一区卜| 热re99久久精品国产66热6| 午夜影院在线不卡| 精品国产一区二区三区久久久樱花| 成人手机av| 黑人猛操日本美女一级片| 欧美 亚洲 国产 日韩一| 亚洲精品一区蜜桃| 成人亚洲欧美一区二区av| 日韩一本色道免费dvd| 免费人妻精品一区二区三区视频| 男人操女人黄网站| 亚洲综合色惰| 亚洲精品国产一区二区精华液| 欧美日韩视频高清一区二区三区二| 春色校园在线视频观看| 久久久久国产精品人妻一区二区| 国产 精品1| 亚洲四区av| a 毛片基地| 黄网站色视频无遮挡免费观看| 另类精品久久| 午夜精品国产一区二区电影| 亚洲伊人色综图| av线在线观看网站| 国产色婷婷99| 在线观看一区二区三区激情| 久久毛片免费看一区二区三区| 国产成人一区二区在线| 中文字幕人妻丝袜制服| 国产一区亚洲一区在线观看| 亚洲国产精品成人久久小说| 久久精品熟女亚洲av麻豆精品| 男女免费视频国产| 欧美中文综合在线视频| 午夜激情av网站| 久久狼人影院| 制服丝袜香蕉在线| 一边亲一边摸免费视频| 少妇人妻精品综合一区二区| 最近最新中文字幕免费大全7| 一区二区三区四区激情视频| 亚洲图色成人| 国产精品一区二区在线观看99| 精品国产乱码久久久久久小说| xxxhd国产人妻xxx| 精品99又大又爽又粗少妇毛片| 高清视频免费观看一区二区| 欧美激情高清一区二区三区 | 亚洲av男天堂| 久久久久久久久免费视频了| 国产精品久久久久成人av| 激情视频va一区二区三区| 婷婷色av中文字幕| av线在线观看网站| 欧美+日韩+精品| 欧美xxⅹ黑人| 又大又黄又爽视频免费| 欧美另类一区| 9191精品国产免费久久| 最近的中文字幕免费完整| 国产爽快片一区二区三区| 午夜影院在线不卡| 欧美老熟妇乱子伦牲交| videosex国产| 永久网站在线| 国产精品成人在线| xxx大片免费视频| 午夜免费鲁丝| 超碰成人久久| 成年人午夜在线观看视频| 性色av一级| av在线app专区| 精品一品国产午夜福利视频| 18禁国产床啪视频网站| 国产白丝娇喘喷水9色精品| 亚洲av国产av综合av卡| 街头女战士在线观看网站| 日本黄色日本黄色录像| 免费高清在线观看日韩| 日本av免费视频播放| 欧美国产精品一级二级三级| 午夜老司机福利剧场| 欧美国产精品va在线观看不卡| 人妻人人澡人人爽人人| 高清黄色对白视频在线免费看| 人成视频在线观看免费观看| 国产精品国产三级专区第一集| 国产精品无大码| 国产成人91sexporn| 成人国产av品久久久| 久久久久久人妻| 涩涩av久久男人的天堂| 欧美97在线视频| 亚洲,欧美精品.| 久久鲁丝午夜福利片| 大陆偷拍与自拍| 一级,二级,三级黄色视频| 男女啪啪激烈高潮av片| 搡女人真爽免费视频火全软件| xxx大片免费视频| 亚洲国产精品一区三区| 午夜福利视频精品| 99久国产av精品国产电影| 国产精品国产三级国产专区5o| 亚洲精品第二区| 久久久久久人妻| 狠狠精品人妻久久久久久综合| 中文字幕亚洲精品专区| 91成人精品电影| 赤兔流量卡办理| 99国产精品免费福利视频| 十八禁网站网址无遮挡| 午夜福利在线免费观看网站| 午夜福利网站1000一区二区三区| 天堂中文最新版在线下载| 国产精品无大码| 美女国产视频在线观看| 国产亚洲欧美精品永久| 少妇猛男粗大的猛烈进出视频| 国产精品99久久99久久久不卡 | 亚洲精品国产一区二区精华液| 爱豆传媒免费全集在线观看| 又粗又硬又长又爽又黄的视频| 国产激情久久老熟女| 欧美精品av麻豆av| 午夜激情久久久久久久| 亚洲国产欧美在线一区| 国产视频首页在线观看| 青春草亚洲视频在线观看| 在线观看国产h片| 亚洲精品av麻豆狂野| 中文字幕精品免费在线观看视频| 亚洲欧美精品综合一区二区三区 | 午夜福利乱码中文字幕| 国产又爽黄色视频| 亚洲精品一区蜜桃| 欧美97在线视频| 国产男女超爽视频在线观看| 午夜福利在线观看免费完整高清在| 亚洲第一区二区三区不卡| 亚洲精品日韩在线中文字幕| 欧美精品一区二区免费开放| 国产精品av久久久久免费| 你懂的网址亚洲精品在线观看| 99国产综合亚洲精品| 亚洲激情五月婷婷啪啪| 亚洲国产av新网站| 亚洲情色 制服丝袜| av电影中文网址| 国产日韩欧美视频二区| 韩国精品一区二区三区| 国产成人精品一,二区| 国产av一区二区精品久久| 亚洲成av片中文字幕在线观看 | 少妇被粗大猛烈的视频| 日韩一卡2卡3卡4卡2021年| av网站在线播放免费| 一级毛片电影观看| 亚洲成人手机| 热99国产精品久久久久久7| 国产乱来视频区| 99久久人妻综合| av在线app专区| 18禁国产床啪视频网站| 国产成人精品福利久久| 叶爱在线成人免费视频播放| 你懂的网址亚洲精品在线观看| 国产黄频视频在线观看| 爱豆传媒免费全集在线观看| 色哟哟·www| 男女国产视频网站| 黄片播放在线免费| 夜夜骑夜夜射夜夜干| 欧美日韩精品网址| 成人二区视频| 亚洲精品一区蜜桃| 久久毛片免费看一区二区三区| 日韩熟女老妇一区二区性免费视频| 视频在线观看一区二区三区| 国产黄频视频在线观看| 十分钟在线观看高清视频www| 天美传媒精品一区二区| 欧美日韩国产mv在线观看视频| 午夜激情av网站| 亚洲国产欧美日韩在线播放| 日日啪夜夜爽| 久热这里只有精品99| 捣出白浆h1v1| 欧美精品一区二区大全| 视频在线观看一区二区三区| 国产精品秋霞免费鲁丝片| 国产精品 欧美亚洲| 卡戴珊不雅视频在线播放| 国产精品国产三级国产专区5o| 久久精品国产亚洲av天美| 巨乳人妻的诱惑在线观看| 熟女av电影| 一级毛片黄色毛片免费观看视频| 只有这里有精品99| 69精品国产乱码久久久| 各种免费的搞黄视频| 成人影院久久| 欧美人与性动交α欧美软件| 精品少妇一区二区三区视频日本电影 | 亚洲国产成人一精品久久久| 免费黄频网站在线观看国产| 欧美激情 高清一区二区三区| 久久久国产精品麻豆| 高清黄色对白视频在线免费看| 久久99蜜桃精品久久| 成人国产麻豆网| 七月丁香在线播放| 大码成人一级视频| 久久韩国三级中文字幕| 少妇人妻精品综合一区二区| 如何舔出高潮| 午夜免费观看性视频| 一级毛片黄色毛片免费观看视频| 久久99蜜桃精品久久| 亚洲一级一片aⅴ在线观看| 久久ye,这里只有精品| 麻豆精品久久久久久蜜桃| 两个人免费观看高清视频| 免费少妇av软件| 人人妻人人澡人人爽人人夜夜| 天天躁夜夜躁狠狠久久av| 精品久久久久久电影网| 欧美中文综合在线视频| 国产成人av激情在线播放| 人妻少妇偷人精品九色| 女人被躁到高潮嗷嗷叫费观| 成人午夜精彩视频在线观看| 成人亚洲欧美一区二区av| 免费在线观看黄色视频的| av网站在线播放免费| 午夜福利视频精品| 国产精品久久久久久精品电影小说| 乱人伦中国视频| 午夜免费男女啪啪视频观看| 国产熟女欧美一区二区| 免费在线观看黄色视频的| 亚洲欧洲精品一区二区精品久久久 | 亚洲精品美女久久久久99蜜臀 | 国产乱来视频区| 久久这里有精品视频免费| 亚洲av欧美aⅴ国产| 久久这里只有精品19| 飞空精品影院首页| 欧美97在线视频| 国产成人精品在线电影| 亚洲美女黄色视频免费看| 精品久久蜜臀av无| 91精品伊人久久大香线蕉| 超碰97精品在线观看| 亚洲精品久久成人aⅴ小说| av.在线天堂| 一区在线观看完整版| 黑人欧美特级aaaaaa片| 男人爽女人下面视频在线观看| 国产精品麻豆人妻色哟哟久久| 女人久久www免费人成看片| 最新的欧美精品一区二区| 大码成人一级视频| 丰满少妇做爰视频| 天天躁夜夜躁狠狠躁躁| 中文字幕人妻丝袜一区二区 | 熟女av电影| 中文字幕另类日韩欧美亚洲嫩草| 欧美日韩精品成人综合77777| 大片电影免费在线观看免费| 丝袜美足系列| 欧美成人精品欧美一级黄| 两个人免费观看高清视频| 男女免费视频国产| 女人精品久久久久毛片| 天天躁日日躁夜夜躁夜夜| 亚洲伊人色综图| 90打野战视频偷拍视频| 国产精品一国产av| 国产激情久久老熟女| 一级毛片我不卡| 晚上一个人看的免费电影| 午夜免费男女啪啪视频观看| 亚洲精品自拍成人| 亚洲精品国产一区二区精华液| 欧美 亚洲 国产 日韩一| 一个人免费看片子| 国产熟女欧美一区二区| 五月伊人婷婷丁香| 黑人猛操日本美女一级片| a 毛片基地| 考比视频在线观看| 午夜福利网站1000一区二区三区| 777米奇影视久久| 一区二区三区四区激情视频| 三级国产精品片| 亚洲三级黄色毛片| 叶爱在线成人免费视频播放| 91国产中文字幕| 制服人妻中文乱码| 热re99久久精品国产66热6| 久久久欧美国产精品| 亚洲一区二区三区欧美精品| 亚洲欧洲日产国产| av不卡在线播放| 国产免费又黄又爽又色| 99国产综合亚洲精品| 国产一区二区在线观看av| 国产成人a∨麻豆精品| 在线亚洲精品国产二区图片欧美| 女性生殖器流出的白浆| av福利片在线| 黄色毛片三级朝国网站| 新久久久久国产一级毛片| 久久国产精品大桥未久av| 亚洲av成人精品一二三区| 欧美日韩视频高清一区二区三区二| 国产无遮挡羞羞视频在线观看| 久久精品aⅴ一区二区三区四区 | 国产国语露脸激情在线看| 精品第一国产精品| av女优亚洲男人天堂| 亚洲欧美色中文字幕在线| 亚洲欧美精品自产自拍| 久久久久国产一级毛片高清牌| 免费在线观看视频国产中文字幕亚洲 | 午夜激情久久久久久久| 精品国产国语对白av| 十八禁高潮呻吟视频| 日韩,欧美,国产一区二区三区| 国产精品无大码| 久久久久久久精品精品| 999精品在线视频| 欧美激情极品国产一区二区三区| 欧美日韩视频精品一区| 亚洲色图综合在线观看| 性色av一级| 国产免费现黄频在线看| 波多野结衣一区麻豆| 毛片一级片免费看久久久久| 婷婷色av中文字幕| 人妻一区二区av| 国产精品成人在线| 久久99热这里只频精品6学生| 久热久热在线精品观看| 在线天堂中文资源库| 国产欧美日韩综合在线一区二区| 免费观看性生交大片5| 中文字幕亚洲精品专区| 免费大片黄手机在线观看| 日本wwww免费看| 咕卡用的链子| 国产xxxxx性猛交| 大型av网站在线播放| 在线播放国产精品三级| 亚洲人成网站在线播放欧美日韩| 80岁老熟妇乱子伦牲交| 五月开心婷婷网| 操美女的视频在线观看| 黄色a级毛片大全视频| 九色亚洲精品在线播放| 亚洲激情在线av| 桃红色精品国产亚洲av| 亚洲色图综合在线观看| 嫩草影视91久久| 交换朋友夫妻互换小说| xxx96com| 自线自在国产av| 少妇裸体淫交视频免费看高清 | 人成视频在线观看免费观看| 久久久久精品国产欧美久久久| 一级黄色大片毛片| 天堂动漫精品| 欧美日韩福利视频一区二区| 一区二区三区国产精品乱码| 波多野结衣av一区二区av| 亚洲一码二码三码区别大吗| 欧美中文综合在线视频| 国产亚洲精品久久久久久毛片| 欧美黑人欧美精品刺激| 精品国产一区二区久久| 精品国产国语对白av| 美女午夜性视频免费| 在线看a的网站| 伦理电影免费视频| 不卡一级毛片| 五月开心婷婷网| 女人被狂操c到高潮| 久久精品亚洲熟妇少妇任你| 在线观看免费视频网站a站| 在线观看午夜福利视频| av国产精品久久久久影院| 久久欧美精品欧美久久欧美| 亚洲欧美一区二区三区黑人| 免费久久久久久久精品成人欧美视频| 久久国产精品影院| 女生性感内裤真人,穿戴方法视频| 免费不卡黄色视频| 免费在线观看视频国产中文字幕亚洲| 久久精品国产亚洲av香蕉五月| 桃红色精品国产亚洲av| 亚洲 欧美一区二区三区| 久久久国产欧美日韩av| 日韩av在线大香蕉| 成人18禁高潮啪啪吃奶动态图| 欧美一区二区精品小视频在线| 久久人妻福利社区极品人妻图片| 狂野欧美激情性xxxx| 久热爱精品视频在线9| 亚洲精品美女久久久久99蜜臀| 在线观看免费午夜福利视频| 国产亚洲精品综合一区在线观看 | 午夜福利一区二区在线看| 国产欧美日韩综合在线一区二区| 老熟妇乱子伦视频在线观看| 国产黄a三级三级三级人| 国产亚洲欧美98| 黄网站色视频无遮挡免费观看| 久久久国产精品麻豆| 国产亚洲欧美98| 久久国产精品影院| 欧美日本中文国产一区发布| 色综合欧美亚洲国产小说| 中文亚洲av片在线观看爽| 国产午夜精品久久久久久| 一级片'在线观看视频| 欧美日韩亚洲高清精品| 少妇裸体淫交视频免费看高清 | 嫁个100分男人电影在线观看| 久久 成人 亚洲| 一区二区日韩欧美中文字幕| 无人区码免费观看不卡| 9热在线视频观看99| 色综合站精品国产| 国内久久婷婷六月综合欲色啪| 国产高清视频在线播放一区| 日本a在线网址| 50天的宝宝边吃奶边哭怎么回事| 国产精品久久久av美女十八| 中文欧美无线码| 在线十欧美十亚洲十日本专区| 女警被强在线播放| 两性夫妻黄色片| 久久久久国产一级毛片高清牌| 搡老乐熟女国产| 国产1区2区3区精品| 午夜免费成人在线视频| 性色av乱码一区二区三区2| 欧美大码av| 久久久国产欧美日韩av| 国产精华一区二区三区| 一区二区三区国产精品乱码| 日韩欧美一区二区三区在线观看| 热99re8久久精品国产| 一边摸一边做爽爽视频免费| 亚洲性夜色夜夜综合| 国产高清国产精品国产三级| 神马国产精品三级电影在线观看 | 一边摸一边抽搐一进一小说| 亚洲国产欧美网| 18禁国产床啪视频网站| 性少妇av在线| 变态另类成人亚洲欧美熟女 | 国产精品一区二区三区四区久久 | 99国产精品一区二区三区| 丰满的人妻完整版| 亚洲免费av在线视频| 久久久久久免费高清国产稀缺| 免费女性裸体啪啪无遮挡网站| 黄色a级毛片大全视频| 在线观看一区二区三区激情| 国产熟女午夜一区二区三区| 国产伦一二天堂av在线观看| 亚洲第一青青草原| 中出人妻视频一区二区| 久久人人97超碰香蕉20202| 一级,二级,三级黄色视频| 99久久99久久久精品蜜桃| 国产精品二区激情视频| 精品久久久久久成人av| 神马国产精品三级电影在线观看 | 成人手机av| 不卡av一区二区三区| 色综合站精品国产| www.熟女人妻精品国产| 最好的美女福利视频网| 成人黄色视频免费在线看| 老司机在亚洲福利影院| 叶爱在线成人免费视频播放| 91老司机精品| 国产三级在线视频| 别揉我奶头~嗯~啊~动态视频| 午夜激情av网站| 香蕉丝袜av| 国产1区2区3区精品| 亚洲成人国产一区在线观看| 精品日产1卡2卡| 日韩免费av在线播放| a在线观看视频网站| 两性午夜刺激爽爽歪歪视频在线观看 | 一边摸一边抽搐一进一出视频| 亚洲精品美女久久av网站| 久久久久精品国产欧美久久久| 亚洲国产精品sss在线观看 | 50天的宝宝边吃奶边哭怎么回事| 国产又爽黄色视频| 美女 人体艺术 gogo| 欧美日韩黄片免| 一级,二级,三级黄色视频| 高清欧美精品videossex| 国产精品日韩av在线免费观看 | 首页视频小说图片口味搜索| 超碰97精品在线观看| 色精品久久人妻99蜜桃| 欧美 亚洲 国产 日韩一| 视频区欧美日本亚洲| 91老司机精品| 男女做爰动态图高潮gif福利片 | 19禁男女啪啪无遮挡网站| 国产高清国产精品国产三级| avwww免费| 狂野欧美激情性xxxx| 国产亚洲精品第一综合不卡| 18禁黄网站禁片午夜丰满| 成人特级黄色片久久久久久久| 俄罗斯特黄特色一大片| 精品久久久久久久久久免费视频 | 99精国产麻豆久久婷婷| 看免费av毛片| 91麻豆精品激情在线观看国产 | 日本五十路高清| 一级作爱视频免费观看| 国产精品99久久99久久久不卡| 中文字幕人妻丝袜制服| 在线国产一区二区在线| 亚洲精品美女久久av网站| 国产人伦9x9x在线观看| 一区二区日韩欧美中文字幕| 99久久综合精品五月天人人| 乱人伦中国视频| 欧美日韩一级在线毛片| 12—13女人毛片做爰片一| 国产精华一区二区三区| 欧美精品亚洲一区二区| 国产成人精品无人区| 中文字幕精品免费在线观看视频| 露出奶头的视频| 一级毛片高清免费大全| 欧美激情极品国产一区二区三区| 人人妻人人添人人爽欧美一区卜| 午夜免费观看网址| 色尼玛亚洲综合影院| 精品一区二区三卡| 狂野欧美激情性xxxx| 欧美日本亚洲视频在线播放| 精品久久久久久电影网| 韩国精品一区二区三区| 一级作爱视频免费观看| 久久久久久免费高清国产稀缺| 久久久久九九精品影院| 一级黄色大片毛片| 十八禁人妻一区二区| 国产视频一区二区在线看| av网站免费在线观看视频| 久久久久国产精品人妻aⅴ院| av在线天堂中文字幕 | 麻豆成人av在线观看| 久久精品91无色码中文字幕| 水蜜桃什么品种好| 高清黄色对白视频在线免费看| 不卡av一区二区三区| 亚洲av第一区精品v没综合| 老司机亚洲免费影院| 男女之事视频高清在线观看| 两性午夜刺激爽爽歪歪视频在线观看 | 侵犯人妻中文字幕一二三四区| 国产亚洲精品久久久久5区| 午夜福利影视在线免费观看| 叶爱在线成人免费视频播放| 在线观看舔阴道视频| av在线天堂中文字幕 | 日韩免费av在线播放| 人人妻人人澡人人看| 97碰自拍视频| 日韩一卡2卡3卡4卡2021年| 免费看a级黄色片| 欧美丝袜亚洲另类 | 伊人久久大香线蕉亚洲五| 麻豆国产av国片精品| 三上悠亚av全集在线观看| 精品熟女少妇八av免费久了| 日韩免费av在线播放| 很黄的视频免费| 新久久久久国产一级毛片| 少妇的丰满在线观看| 久久国产精品男人的天堂亚洲| 成人亚洲精品av一区二区 | 成年女人毛片免费观看观看9| 天天影视国产精品| 亚洲avbb在线观看| 他把我摸到了高潮在线观看| 久久青草综合色| 国产免费现黄频在线看| 久久精品亚洲精品国产色婷小说| 性色av乱码一区二区三区2| 国产免费男女视频| 国产成人精品无人区| 国产精品免费视频内射| 大香蕉久久成人网| 国产高清视频在线播放一区|