• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Transforming Hand Drawn Wireframes into Front-End Code with Deep Learning

    2022-11-11 10:44:16SamanRiazAliArshadShahabBandandAmirMosavi
    Computers Materials&Continua 2022年9期

    Saman Riaz,Ali Arshad,Shahab S.Bandand Amir Mosavi

    1Department of Computer Science,National University of Technology,Islamabad,44000,Pakistan

    2Department of Computer Science,Institute of Space Technology,Islamabad,44000,Pakistan

    3Future Technology Research Center,National Yunlin University of Science and Technology,Douliu,64002,Yunlin,Taiwan

    4Faculty of Civil Engineering,Technische Universitat Dresden,Dresden 01069,Germany

    Abstract: The way towards generating a website front end involves a designer settling on an idea for what kind of layout they want the website to have,then proceeding to plan and implement each aspect one by one until they have converted what they initially laid out into its Html front end form,this process can take a considerable time,especially considering the first draft of the design is traditionally never the final one.This process can take up a large amount of resource real estate,and as we have laid out in this paper,by using a Model consisting of various Neural Networks trained on a custom dataset.It can be automated into assisting designers,allowing them to focus on the other more complicated parts of the system they are designing by quickly generating what would rather be straightforward busywork.Over the past 20 years,the boom in how much the internet is used and the sheer volume of pages on it demands a high level of work and time to create them.For the efficiency of the process,we proposed a multi-model-based architecture on image captioning,consisting of Convolutional neural network(CNN)and Long short-term memory(LSTM)models.Our proposed approach trained on our custom-made database can be automated into assisting designers,allowing them to focus on the other more complicated part of the system.We trained our model in several batches over a custom-made dataset consisting of over 6300 files and were finally able to achieve a Bilingual Evaluation Understudy (BLEU)score for a batch of 50 hand-drawn images at 87.86%

    Keywords: Deep learning; wireframes; front-end; low fidelity; high fidelity;design process;html;computer vision;dsl

    1 Introduction

    Nowadays,transforming hand-drawn wireframes into front-end code is created by a designer into computer code is a typical customized software, websites, and mobile apps [1-3], which serves as a rough idea of how we want the final structure to be.A key issue with this is that because sites are built of dozens if not thousands of different pages,having to create and design a sketch then converting that into functioning code can be extremely costly and time-consuming,as designers need to go through one by one and implement each required aspect of the final structure[1,4].All this time adds up and takes up a significant amount of development time for the projects, time that could be better used elsewhere,such as the more technical aspects of the work.

    In our proposed work,the process of converting the wireframes to code involves passing the design to a developer/UI designer, having them implement the boilerplate graphical user interface (GUI)[1,5,6],and then reiterate this product until they reach a stage that they or the client are satisfied with.

    Deep learning changed front-end development by increasing prototypes speed and lowering the barrier for building software.In this paper, we tackle the issue and aim to facilitate the needs of designers by saving them a significant amount of time.Using deep learning techniques[2-3,7-9]would enable the automation of the creation process by simply letting the designers take their sketches and manipulate them using the model into HTML code.Using Neural Networks,we can employ them to learn the underlying patterns and context of elements on a sketch and convert them into front-end Html code [1,4], which they could then edit, customize or implement according to their needs.The objectives of our system are such that:

    ? Saves time:Removing the need for a large amount of busy work frees the designers to work on other more complicated or involved portions of their work as now the time to sketch to code conversation does not need to be painstakingly done over and over.

    ? Focus: Designing front-ends can sometimes be tedious and repetitive work,preventing developers from focusing on actual core logic.This system will help them focus on actual problems.

    ? Flexibility: This will allow us to quickly transition between different designs to test out what works best with the desired application,spending less time and resources with each iteration.

    ? Accuracy: With the implementation of our proposed technique, the system can produce an accurate extrapolation of the given sketches into their proper code form with an average of 87.86%accuracy.

    The contribution of our proposed system, we found the best approach is through the usage of multi-Model deep learning[2-3,7-9]and computer vision[7-9]techniques to train a system to handle the vast number of variations in a hand-drawn wireframe.We believe that through a rigorous training process and implementation of these concepts, we can create a model capable of servicing various users’needs and generating a front end with a high amount of accuracy.

    There were some major challenges in the creation of such a system:

    ? Dataset Disparity for a deep learning model.

    ? A large amount of time and resources are required in the creation of a new custom dataset.

    ? Hardware Limitations due to costly training times.

    ? Dividing the issue into sub problems,namely computer vision and language modeling problems.

    ? Getting each element to be recognized by the model,and correctly translated.

    2 Backgrounds

    2.1 Sketches

    Sketching is a Hand-drawn mockup or User interface design shown in Fig.1;it is the basis of a user interface before prototyping[10,11],and coding.Sketching[12]comes after the idea of a website or any application before building screen mockups in a website wireframe tool.Sketching refers to the act of working through all the possible ways you could make your idea into an interface reality[12].At the end of the UI sketching process,you should know that you have figured out the absolute best way to bring your product to the screen.However,you choose to sketch your initial User Interface ideas;keeping it simple and not spending too much time on it is essential.

    Figure 1:Representation of a user interface designed for a website

    2.2 Wireframes

    A wireframe[13,14]is a simple representation of a products design.They represent a way to get ideas down quickly.They help designers focus on the more functional aspect before moving on to the finer details.They are generally two types of wireframes:Low-fidelity wireframes[15]and Highfidelity wireframes [16].A low-fidelity wireframe is a simple representation shown in Fig.2a.It can be a drawing on a whiteboard,a rough sketch drawn on pieces of paper,just as long as it can properly communicate the purpose of what we want to create.High-fidelity wireframes are more complete representations of the end product,which is shown in Fig.2b.They help to communicate how aesthetic features support essential functionalities.

    Figure 2:Representation of low fidelity wireframes(a)and high fidelity wireframes(b)

    2.3 Frontend

    Front-end design consists of user experience elements on the web page or application,including menus, pages, buttons, icons, links, graphics, etc.The front-end is a combination of the following technologies such as HTML[17],CSS[18],and Javascript[19].

    3 Literature Review

    In this section of the paper,we explore the current studies and research done in the field of sketch to code conversion.

    Beltramelli proposed pix2code [4], which managed to implement, which managed to convert high-fidelity wireframes into code.They trained their model with stochastic gradient descent[20]to simultaneously learn to model sequences and Spatio-temporal visual features[21]to generate variablelength strings of tokens from a single GUI image as input.It was designed to process the input data;the Author model learns from the pixel values of the input image alone.Pix2code converted screenshots to code,but it cannot still convert hand-drawn wireframes into code.At the same time,our proposed method converted hand-drawn sketches into code.

    Alexander Robinson proposed Sketch2code [1] by which attempts to solve the problem of translating wireframe sketches into code using deep learning [22] that showed considerable success over classical techniques when applied to other domains,particularly in vision problems.

    DeepCoder [23], a system created to generate computer programs by leveraging statistical predictions to augment traditional search techniques done by MatejBalog, Alexander L.Gaunt, Marc Brockschmidt, Sebastian Nowozin, and Daniel Tarlow.In another work by Gaunt et al., the generation of source code is enabled by learning the relationships between input-output examples via differentiable interpreters.Program generation like this is a very active research field but program generation from taking visual inputs is still an unexplored area that our research paper attempts to explore.

    Tuan Anh Nguyen and ChristophCsallner on Reverse Engineering Mobile Application User Interfaces(REMAUI)[24].REMAUI identifies user interface elements such as images,texts,containers,and lists,via computer vision[25]and optical character recognition(OCR)[26]techniques.This is a good example of how these types applications take a wireframe or high-fidelity image and translate them into code.These applications with structure also solve all the associated styles with it,but it faces the problem of adding new examples,which is challenging and time-consuming.The problem is made worse by the inclusion of style in the designs,leading to a significantly higher variation than the digital drawings in the low fidelity designs[1].

    SILK [27] by James A.Landay (Human-Computer Interaction Institute School of Computer Science Carnegie Mellon University).SILK is an interactive tool that allows designers to sketch and an electronic pad quickly.It then retains the“sketchy”look of the components.The system facilitates quick prototyping of interface ideas through the use of gestures for sketch creation and storyboards.

    It uses a combination of computer vision [7-9] techniques to classify shapes drawn on a digital surface into predefined components such as corner detection and line detection,classify shapes and application elements such as buttons,text,textfield,box,etc.The main advantage of SILK[27]over paper sketches is that it allows the storyboards to come alive and permits the designer or test subjects to exercise the interface in this early,sketchy state.

    While the concept of turning wireframes into code is not new,as with the advancement of machine learning, more work has been put into it, but it still runs into some complex challenges.Thus, our primary analysis focused on two key papers,Sketch2code[1]and Pix2code[4],each concerned with a specific conversion aspect.

    We have focused on two main research areas: the first one in turning high fidelity wireframes such as digital mockups into code and secondly converting low fidelity drawings into code.Both these studies have one thing in common: the translation of wireframes designs into a coded form.The significant difference is the fidelity of the designs being translated.

    From our examination of the current work done, we saw that there was a limited number of elements considered in the research, which we believe could be improved upon, along with this the primary dataset contained GUI images and their associated code,so we decided to make alterations within the primary dataset to make it look more hand-drawn.To this end, we created a model with various optimizations to ensure better accuracy and training times,along with an increased amount of considered elements in its conversion from wireframes into code.

    4 Techniques

    4.1 Computer Vision Techniques

    To build a system that could detect the various aspects of our sketches, such as the lines and edges, we needed to use computer vision techniques [7-9].Computer vision is how a computer can understand images or videos;it relies on acquiring,processing,and analyzing digital images and using information supplied to it,to create a context in which to break down said images.

    4.1.1 Edge Detection

    To make our system work,we need to consider how we go about detecting the various elements in our wireframe[13]sketches.To this end,when we draw a wireframe symbol consisting of straight or sometimes rounded edges,the CNN will detect these edges effectively given enough sample size.

    There are various techniques commonly used to detect edges:

    ? The Sobel operator [28] computes the gradient map of an image and then yields edges by thresholding the gradient map.

    ? The Canny operator [29] extends the Sobel operators, which adds Gaussian smoothing as a preprocessing step and uses the bi-threshold to get the edges.

    ? Richer Convolutional Features[30]is a state-of-the-art deep learning approach using convolutional networks.

    4.1.2 Segmentation

    The sketches that one would create many types of elements, for example, buttons, image boxes,radio buttons,checkboxes,text boxes,etc.To make sure these various elements are detected adequately to be converted into code,we need a system to segment[31]each part of the sketch and classify them for our system to convert them to their correct forms.

    The approach we researched consisted of using a U-Net Model[32]that is trained to segment a hand-drawn image by color-based segmentation [33].It derives its name from how its architecture,when visualized, seems to lay out what it is.It derives its name from how its architecture, when visualized,seems to lay out what it is can be said as the shape U.

    4.1.3 Data Augmentation

    One of the key hurdles for our work was the lack of available data; to this end, we needed to employ the concept of data augmentation[34].Data augmentation is the technique of increasing the amount of data available by slightly modifying different copies of the existing data and turning them into new data.

    Our methods of data augmentation involved manipulating our current hand-drawn images by Altering the scale, Flipping and Rotating, Changing Brightness and Contrast and Adding Noise.Through this augmentation,we were able to achieve a certain amount of variation within the dataset,mainly used to counter overfitting problems that arise from the use of our synthesized dataset.

    4.2 Scripting Techniques

    To increase the sample size of the primary dataset,we finally decided to employ a custom script that would directly alter the CSS elements present in the GUI images.We tried to introduce various human elements and variations to the images like skewed lines, fonts that resemble human writing,rounded off edges,clear white background with the elements drawn in black,etc.We were successfully able to replicate a hand-drawn feeling to the primary dataset,though produced synthetically.

    4.3 Deep Learning Techniques

    Machine learning serves as the basis for our entire system to work.It is a greatly emerging and powerful tool used to give computers the ability to learn and adapt to serve purposes without explicitly programming each function.It achieves this by being fed a large amount of data and “teaching”it how to analyze and use that data to then work on newer information for a specific task.We chose machine learning as it is excellent for classification and detection tasks,along with helping to automate busywork significantly.In this section,we talk about the essential techniques we explored and used in our approach.

    4.3.1 Convolutional Neural Networks

    Convolutional Neural Networks [30,35] are deep learning [3] algorithms that take input images and assign a level of weight and basis to them.Using those weights can better differentiate one image from another.

    CNN’s use raw pixel data as input when incorporated into image-based projects,and they consist of input,output,and hidden layers shown in Fig.3,and these hidden layers consist of Convolutional layers,Pooling layers,and Fully connected layers.The way we have incorporated CNNs in our work is an Encoder.This encoder used as means through which the model encodes, and understands the elements present in the hand-drawn image.With this,we can classify various types of elements present in a Wireframe[13]Image.

    4.3.2 Recurrent Neural Networks

    RNNs are a robust form of neural network that uses their internal state (memory)to process a variable length of sequences of inputs;this makes them great for applications involving handwriting recognition.There are two of the types of RNNs that are made use of are Long short-term memory(LSTM)[36]and Gated recurrent units(GRUs)are a gated mechanism in RNNs;they lack an output gate and thus have fewer overall parameters than an LSTM.They serve well as encoders and decoders[37],as they provide context to the images in the form of a sequence of tokens.In our system we used GRU as a decoder.

    Figure 3:The architecture of Convolutional Neural Network

    4.3.3 Fine Tuning

    All image tensors[38]were compressed into simple images and were converted into low-resolution images to save memory.These low-resolution images were then converted again into image tensors for further resource-saving.Although the model has been relatively optimized since it started working,it is costly to Pre-Process,so to evaluate it,we had to reduce the sample size.In addition to this,to train such a complicated Model with our Improvements on the total sample size,we used a Cloud Server[32](Google Cloud Services)to train it.

    5 Methodology

    To conduct an applied research study and develop methods that convert a Hand Drawn Wireframe into a working HTML code with a Hand Drawn Wireframe into a working HTML code with a Hand Drawn Wireframe into a working HTML code with Deep Learning.We used a primary dataset acquired from the research conducted by Tony Beltramelli[39],which contained over 1750 Images with relation to their appropriate code files.We used this primary dataset [39] as a basis for our dataset that we later improved upon, and trained over a series of Neural Networks [40] based on an Image Captioning Architecture.With the help of this approach,we were able to convert a given Hand Drawn Image of a Wireframe[1,15,16]into its HTML[13]file.

    5.1 Dataset

    Our approach required a larger dataset consisting of hand-drawn wireframes and their associated code.

    We used Tony Beltramelli’s research dataset [39] as our primary dataset, which contained over 1750 GUI Images and their associated Code Files.However,since we required Hand Drawn Images,we had to take specific steps to achieve a modified Dataset that our approach used effectively.Our modified dataset contains Images that are limited to look as close as possible to a hand-drawn image shown in Fig.4 and has a sample size of 6300.We also altered the images and the code files with the addition of newer elements which is shown in Fig.5.

    Figure 4:Wireframe image from our dataset

    Figure 5:Elements key for our wireframe images

    5.2 Framework

    Our expected input image was a white background with elements drawn in a darker color fed into the Framework.The front end was intended to make it easier to feed the input from the User.The image was then processed into the Model, where Model tries to learn elements present in the image and the context that each element has,including its sequence and structure.Our proposed Framework shows in Fig.6 how a wireframe image is passed into our Deep Learning model and how it is eventually processed into our required output.The output of the Model was a sequence of DSL[1,4]tokens shown in Fig.6, which is Twitter Bootstrap [41], the reason being the limit on the sequence of GRU [5,42]tokens.These DSL[1,4]tokens were then fed into a custom-made compiler that was finally converted into an HTML[13]file,the expected output from our Framework.

    Figure 6:Flowchart of our proposed approach

    5.3 Technical Approach

    Our proposed image captioning architecture based model shown in Fig.7.Every image is first encoded with a CNN-based encoder[5]model,and every code part that is first converted into one-hot encoding[43]is then encoded with a language model composed of stacks of GRU[44]layers.These two encoders are then concatenated and are fed into another stack of GRU layers which behave as a decoder[5].It performs token-level modeling by using one-hot encoded vectors[45],eliminating any need for word embedding techniques such as word2vec [20], resulting in costly computations.After which, a softmax layer [43] is used to output tokens based on the DSL [1,4] vocabulary size.Those tokens are then fed into the custom bootstrap HTML compiler, which outputs HTML code.It is created with the help of TensorFlow and Keras[46].

    A rough image along with its associated code is used to train the model.After the model is trained,it gives us the output in the form of tokens that are then passed into the compiler,which generates its pure HTML and CSS shown in Fig.8.

    6 Experiment

    6.1 Experiment Setup

    As stated before,we decided to use Tony Beltramelli’s research dataset[39],which consisted of over 1750 GUI images and their associated code files.However,since our input requirement was a handdrawn image, we required low-fidelity wireframes and associated code files.Since such a convenient dataset was not available in quality and quantity,we decided to alter and make modifications within the primary dataset,to give it a feel as close as possible to a hand-drawn image.We researched various techniques for making this possible;however,we ended up relying on a script that directly alters the CSS properties of the GUI image front ends.The images were altered in a way to produce a handdrawn image shown in Fig.9.

    Figure 7:Architecture of our proposed approach

    Figure 8:HTML/CSS

    The very first step we did was to pre-process our image data.Load images from an npz file extension by converting them to a png format, using the PIL library.After loading the images, we resized them from their former size of 1380×2400 to 256×256.This preprocessing step aimed to reduce the amount of processing power needed to process UINT8 image tensors.The image tensors were converted to npz and fed into the model.The next step was to pre-process the language files of the dataset.A Vocabulary is made,which consists of all the elements that are present within an image.We updated this vocabulary by the addition of a newer element for classification.

    A total of 18 classes were used to predict the labels that are mentioned in Tab.1,using Keras we have a detailed summary of the model.

    Total parameters shown in Tab.2 are 139,723,686; apart from this, all libraries used in building and evaluating this project is as numpy, lorem_text, nltk, argparse, keras.models, keras.callbacks,keras.layers,keras.optimizers,spacy.

    Figure 9:Our hand-drawn image

    Table 1: Vocabulary

    Table 2: Model parameters

    6.2 Performance Measure

    To measure and evaluate the performance of our model,we used the following methods:

    1.BLEU Score:To evaluate and analyze the performance of our proposed method we use to evaluate it using a batch of images on the BLEU, or the Bilingual Evaluation Understudy metric,it is a score for comparing a candidate translation of the text to one or more reference translations,where 1.0 would represent a perfect score whilst a 0.0 would indicate a poor score.It is mathematically expressed as:

    which is the N-gram precision bounded above by the highest count of n-grams in any reference sentence.

    a penalty added for short translations.

    consists of the brevity penalty and the geometric mean of N-gram precisions.

    2.Spacy’s document similarity:Spacy’s similarity refers to the semantic similarity between words,phrases or even between two whole documents.The core logic behind this is to create two representative vectors of two items; using either universal vector created from pre-trained models like word2vec, glov, fasttext, bartetc, spacy similarity is a cosine similarity between the internal vectors for the document spacy creates.We selected five random documents with each predicted and actual translation, and by using spacy’s similarity, we obtained results in Tab.3.

    Table 3: Spacy’s document similarity

    6.3 Results and Analysis

    As shown in Tabs.2-5 and their associated graphs in Figs.10-12, the training process for the model on our altered dataset consisting of over 6300 files shows the gradual decrease in the loss and validation loss values after each batch was trained.We trained our images over four batches to accommodate the complex pre-processing required before our training process.We observed from the results we collected that there were spikes amidst the validation loss values and have concluded that it was caused mainly because the dataset was synthesized to produce a hand-drawn feel and was not hand-drawn apart from the addition of the newer element.

    Table 4: Loss and validation loss

    Table 5: Third and fourth batch epoch,loss,validation loss

    Table 5:Continued

    Figure 10:Graph of loss and validation loss values for the batch(1st,2nd)vs.epoch size

    In Figs.10a and 11a,we can see that our model’s loss score and validation score are moving side by side,but as we progress further in Figs.10b and 11b,the validation score becomes increasingly jagged.By observing the graph and its lines,it is clear that the model is overfitted.Although the newer element we introduced, i.e., hand-drawn images, into the dataset, all the other elements were synthesized to produce the effect of a handmade drawing.This caused our model to learn the synthesized part of the dataset to the extent that it caused negative impact on overall results.

    However, some of the negative impacts were reduced using overfitting reduction techniques,namely by adding Gaussian noise and increasing the dropout of nodes.The model incorporated with these techniques was trained again.We can see in Fig.13,and Tab.6 do somewhat mitigate the negative effects of a synthesized dataset,but ultimately it reduces accuracy.

    Figure 11:Graph of loss and validation loss values for the batch(3rd,4th)vs.epoch size

    Figure 12:Graph of loss and validation loss for all the batches vs.epoch size

    Figure 13:Graph of final loss and validation loss values of each batch in M1 and M2

    Table 6: Comparison of loss values between original model(M1)and our modified model(M2)

    In the end,we obtained an average BLEU score of 0.8785 and received a favorable outcome from spacy’s document similarities.However,these results can be further refined by using a dataset of actual hand-drawn images instead of synthesized images and increasing the quantity of the overall dataset with more significant variation.The model parameters can then be readjusted to coincide with the dataset to produce a better result.

    7 Conclusions

    Our proposed method of Transforming Hand Drawn Wireframes into Front-End Code based on multi-model Deep Learning techniques can successfully translate hand-drawn images based on an established vocabulary with high accuracy and less time if the user adheres to will result in a clean output for an Html front-end page.We enhanced the previous studies on the subject by increasing the sample size of the preliminary dataset and added a newer element for classification and an analysis of its performance.Various optimizations were made in the model to process the complicated data with more ease and greater accuracy than previous implementations.

    Our study into the subject of image conversion and multi-model deep learning techniques also made us aware of the limitations that a project like this has, namely that, to further increase the accuracy of this or a model of similar nature, a far greater dataset of quality and quantity will be required,along with training for much greater variations and elements by using data based on actual drawings that contain noise as well,because as our results concluded there are bound to be limitations to a synthetically produced image.With the work done in this paper multi-model approach,enough groundwork has been established for further extension of this work as deep learning,and computer vision are both emerging fields.

    Author Contributions:S.R.Conceptualization and Methodology;A.A.Software,Writing—review&editing;S.S.funding acquisition;and A.M.Editing.

    Acknowledgement:We would like to thank Sir Syed CASE Institute of Technology for their resources and help throughout both the development of this project,as well as our time gaining the knowledge and tools we would need to succeed in the professional world and Tony Beltramelli for providing us with the base dataset and his precious work on his paper on which we extended our research.

    Funding Statement:The authors received no specific funding for this study.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    美女高潮到喷水免费观看| 丝瓜视频免费看黄片| 国产精品国产av在线观看| 国产在线免费精品| 亚洲人成电影观看| 久久精品国产99精品国产亚洲性色 | av有码第一页| 交换朋友夫妻互换小说| 首页视频小说图片口味搜索| 亚洲精品av麻豆狂野| 久久国产精品影院| 国产男靠女视频免费网站| 成年人免费黄色播放视频| 纵有疾风起免费观看全集完整版| 日韩中文字幕视频在线看片| 亚洲va日本ⅴa欧美va伊人久久| 久久婷婷成人综合色麻豆| 久久精品国产亚洲av香蕉五月 | e午夜精品久久久久久久| 精品国内亚洲2022精品成人 | 悠悠久久av| 久久天堂一区二区三区四区| 精品国产乱码久久久久久男人| 女警被强在线播放| 日日摸夜夜添夜夜添小说| 动漫黄色视频在线观看| 精品人妻熟女毛片av久久网站| 1024视频免费在线观看| 久久人妻av系列| 黄色成人免费大全| 91麻豆精品激情在线观看国产 | 黄色视频,在线免费观看| 别揉我奶头~嗯~啊~动态视频| 伦理电影免费视频| 亚洲成人免费电影在线观看| 久久久精品免费免费高清| 亚洲精品中文字幕一二三四区 | 少妇被粗大的猛进出69影院| 男女边摸边吃奶| 国产亚洲欧美精品永久| 下体分泌物呈黄色| 久久久欧美国产精品| 男女下面插进去视频免费观看| 一本—道久久a久久精品蜜桃钙片| 热re99久久国产66热| 每晚都被弄得嗷嗷叫到高潮| 国产精品二区激情视频| 视频在线观看一区二区三区| 成人18禁高潮啪啪吃奶动态图| 黄色成人免费大全| 国产精品二区激情视频| 欧美日韩视频精品一区| 免费av中文字幕在线| 久久精品国产综合久久久| 韩国精品一区二区三区| 国产精品av久久久久免费| 国产成人一区二区三区免费视频网站| 亚洲成人手机| 国产精品亚洲一级av第二区| 亚洲视频免费观看视频| 夜夜爽天天搞| 在线播放国产精品三级| 国产精品亚洲一级av第二区| 亚洲成人国产一区在线观看| 最黄视频免费看| 在线播放国产精品三级| 黄色 视频免费看| 少妇被粗大的猛进出69影院| 汤姆久久久久久久影院中文字幕| 日本av手机在线免费观看| 女人爽到高潮嗷嗷叫在线视频| 婷婷成人精品国产| 国产午夜精品久久久久久| 成人手机av| 大码成人一级视频| 久热爱精品视频在线9| 精品视频人人做人人爽| 亚洲五月婷婷丁香| 制服人妻中文乱码| 一边摸一边抽搐一进一出视频| 国产又爽黄色视频| 又大又爽又粗| xxxhd国产人妻xxx| 好男人电影高清在线观看| 亚洲av日韩精品久久久久久密| 大香蕉久久成人网| 在线观看免费高清a一片| 亚洲国产成人一精品久久久| 日韩视频一区二区在线观看| 大片电影免费在线观看免费| 侵犯人妻中文字幕一二三四区| 亚洲精品久久成人aⅴ小说| 夜夜爽天天搞| av国产精品久久久久影院| 国产高清视频在线播放一区| 亚洲精品粉嫩美女一区| 亚洲国产av影院在线观看| 99久久99久久久精品蜜桃| 午夜福利视频在线观看免费| 高清黄色对白视频在线免费看| 一级片'在线观看视频| 欧美日韩av久久| 国产精品香港三级国产av潘金莲| 国产黄频视频在线观看| 久久青草综合色| 建设人人有责人人尽责人人享有的| 欧美黄色片欧美黄色片| 精品一区二区三区四区五区乱码| 岛国在线观看网站| 在线永久观看黄色视频| 久久久国产成人免费| 国产成人精品无人区| 十八禁人妻一区二区| 亚洲国产av新网站| 国产亚洲午夜精品一区二区久久| 男男h啪啪无遮挡| 成年版毛片免费区| 男人舔女人的私密视频| 亚洲视频免费观看视频| 人成视频在线观看免费观看| 99国产精品一区二区蜜桃av | 老熟妇仑乱视频hdxx| 天天影视国产精品| 精品一区二区三区视频在线观看免费 | 99久久国产精品久久久| 久久香蕉激情| 操美女的视频在线观看| 99精品久久久久人妻精品| av又黄又爽大尺度在线免费看| 欧美精品av麻豆av| 午夜免费鲁丝| 色播在线永久视频| 高清欧美精品videossex| 亚洲成人免费电影在线观看| 天天影视国产精品| av有码第一页| cao死你这个sao货| 日韩有码中文字幕| 欧美国产精品一级二级三级| 中文字幕高清在线视频| 热99国产精品久久久久久7| 久久久久久人人人人人| 中文字幕人妻丝袜一区二区| 国产单亲对白刺激| 久久久精品区二区三区| 久久久久精品人妻al黑| 免费人妻精品一区二区三区视频| 欧美乱码精品一区二区三区| kizo精华| 国产av又大| 精品少妇内射三级| 欧美av亚洲av综合av国产av| 日本av免费视频播放| netflix在线观看网站| 国产精品久久久av美女十八| 大香蕉久久成人网| 不卡av一区二区三区| avwww免费| 欧美日韩成人在线一区二区| 日本av手机在线免费观看| av天堂在线播放| 纵有疾风起免费观看全集完整版| 丁香六月欧美| 日韩精品免费视频一区二区三区| 亚洲五月婷婷丁香| 男男h啪啪无遮挡| 午夜福利欧美成人| 午夜精品久久久久久毛片777| 丝袜美足系列| 无遮挡黄片免费观看| 交换朋友夫妻互换小说| 啦啦啦中文免费视频观看日本| 亚洲av成人不卡在线观看播放网| 母亲3免费完整高清在线观看| 国产精品一区二区免费欧美| 久久久久久人人人人人| 老鸭窝网址在线观看| 天堂8中文在线网| 91字幕亚洲| 亚洲精品中文字幕一二三四区 | 人人妻人人爽人人添夜夜欢视频| 成人国产av品久久久| 在线看a的网站| 高清在线国产一区| 少妇精品久久久久久久| 免费看十八禁软件| 香蕉久久夜色| 在线天堂中文资源库| 久久性视频一级片| 看免费av毛片| 极品少妇高潮喷水抽搐| 久久国产精品大桥未久av| 日本撒尿小便嘘嘘汇集6| 色婷婷久久久亚洲欧美| 免费高清在线观看日韩| 天堂8中文在线网| 波多野结衣av一区二区av| 亚洲精华国产精华精| 亚洲欧美日韩高清在线视频 | svipshipincom国产片| 老汉色∧v一级毛片| 国内毛片毛片毛片毛片毛片| 黄片播放在线免费| 精品视频人人做人人爽| 亚洲国产av新网站| 国产一区二区三区视频了| 亚洲人成77777在线视频| 2018国产大陆天天弄谢| 91字幕亚洲| cao死你这个sao货| 男男h啪啪无遮挡| 精品人妻在线不人妻| 久9热在线精品视频| 又黄又粗又硬又大视频| 国产精品九九99| 一级片'在线观看视频| 老司机靠b影院| 亚洲七黄色美女视频| 久久99热这里只频精品6学生| 丁香六月天网| 在线观看免费午夜福利视频| 侵犯人妻中文字幕一二三四区| 90打野战视频偷拍视频| 国产日韩一区二区三区精品不卡| 99re在线观看精品视频| 色婷婷av一区二区三区视频| 欧美日韩黄片免| 俄罗斯特黄特色一大片| 久久国产精品人妻蜜桃| 一个人免费看片子| 一区在线观看完整版| 夜夜骑夜夜射夜夜干| 18禁国产床啪视频网站| 欧美一级毛片孕妇| 大型黄色视频在线免费观看| 成人手机av| 超碰成人久久| 性高湖久久久久久久久免费观看| 国产aⅴ精品一区二区三区波| 国产在线一区二区三区精| 老鸭窝网址在线观看| 成人精品一区二区免费| 日本精品一区二区三区蜜桃| 老司机午夜福利在线观看视频 | 老鸭窝网址在线观看| 日日摸夜夜添夜夜添小说| 亚洲五月色婷婷综合| 欧美午夜高清在线| 大香蕉久久成人网| 我的亚洲天堂| 少妇粗大呻吟视频| 91成年电影在线观看| 色婷婷av一区二区三区视频| 亚洲午夜精品一区,二区,三区| 国产欧美日韩一区二区三| 男人操女人黄网站| 美女福利国产在线| 国产日韩欧美亚洲二区| 国产福利在线免费观看视频| 黑人猛操日本美女一级片| 国产精品98久久久久久宅男小说| 一本一本久久a久久精品综合妖精| tocl精华| 啦啦啦在线免费观看视频4| 飞空精品影院首页| 国产一区二区三区在线臀色熟女 | 亚洲av日韩精品久久久久久密| 男女高潮啪啪啪动态图| 日韩视频在线欧美| 国产精品亚洲av一区麻豆| 久久久久久免费高清国产稀缺| 最近最新中文字幕大全免费视频| 精品久久久精品久久久| 亚洲精品粉嫩美女一区| 黄色a级毛片大全视频| 国产又色又爽无遮挡免费看| 成人国语在线视频| 12—13女人毛片做爰片一| av超薄肉色丝袜交足视频| 精品少妇内射三级| 国产精品九九99| 亚洲情色 制服丝袜| 在线观看一区二区三区激情| 国产成人免费观看mmmm| 99久久人妻综合| 高清毛片免费观看视频网站 | 91麻豆av在线| 午夜福利视频在线观看免费| 视频区图区小说| 黄网站色视频无遮挡免费观看| 国产亚洲精品久久久久5区| xxxhd国产人妻xxx| 男女床上黄色一级片免费看| 最新美女视频免费是黄的| tube8黄色片| 99香蕉大伊视频| 国产精品99久久99久久久不卡| 久久国产精品大桥未久av| xxxhd国产人妻xxx| 午夜福利,免费看| 亚洲avbb在线观看| 亚洲国产毛片av蜜桃av| 极品人妻少妇av视频| 免费在线观看影片大全网站| 嫁个100分男人电影在线观看| 在线十欧美十亚洲十日本专区| 视频区欧美日本亚洲| 国产aⅴ精品一区二区三区波| 日韩 欧美 亚洲 中文字幕| 国产成人一区二区三区免费视频网站| 国产欧美日韩一区二区三区在线| 欧美乱妇无乱码| 午夜激情av网站| 大陆偷拍与自拍| 久久青草综合色| 亚洲精品一二三| 久久精品国产99精品国产亚洲性色 | 伊人久久大香线蕉亚洲五| 夜夜爽天天搞| 欧美激情久久久久久爽电影 | 亚洲一区中文字幕在线| 免费观看av网站的网址| 窝窝影院91人妻| 久久精品aⅴ一区二区三区四区| 日本撒尿小便嘘嘘汇集6| 美女视频免费永久观看网站| 久久精品亚洲av国产电影网| 天堂俺去俺来也www色官网| 久久久久久久精品吃奶| 热99国产精品久久久久久7| 90打野战视频偷拍视频| 日日爽夜夜爽网站| 国产在线免费精品| 亚洲 欧美一区二区三区| 精品乱码久久久久久99久播| 男女高潮啪啪啪动态图| 另类亚洲欧美激情| 欧美日韩亚洲国产一区二区在线观看 | 女人爽到高潮嗷嗷叫在线视频| 精品一区二区三卡| 成年人黄色毛片网站| av免费在线观看网站| 另类亚洲欧美激情| 99热网站在线观看| 三上悠亚av全集在线观看| 大香蕉久久成人网| 欧美成人午夜精品| 久久免费观看电影| 成年人午夜在线观看视频| 一个人免费在线观看的高清视频| 午夜福利视频精品| 啦啦啦 在线观看视频| 成人18禁高潮啪啪吃奶动态图| 国产xxxxx性猛交| 亚洲国产av新网站| 99久久人妻综合| 一进一出抽搐动态| 亚洲专区中文字幕在线| 热99re8久久精品国产| 国产成人欧美| 男人操女人黄网站| 丝袜在线中文字幕| 男人操女人黄网站| 在线亚洲精品国产二区图片欧美| 精品熟女少妇八av免费久了| 亚洲专区中文字幕在线| 99精品久久久久人妻精品| 国产亚洲av高清不卡| 国产成人精品久久二区二区91| 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲精品在线美女| 久久精品成人免费网站| 精品福利观看| 日韩人妻精品一区2区三区| 久久精品亚洲精品国产色婷小说| 欧美日韩亚洲综合一区二区三区_| 亚洲国产欧美网| 国产三级黄色录像| 国产免费视频播放在线视频| 成人免费观看视频高清| 一区二区三区国产精品乱码| 国产免费现黄频在线看| 丰满少妇做爰视频| 久久人妻熟女aⅴ| 女人爽到高潮嗷嗷叫在线视频| 国产精品自产拍在线观看55亚洲 | 亚洲伊人色综图| 久久这里只有精品19| 欧美+亚洲+日韩+国产| 18禁黄网站禁片午夜丰满| 欧美 日韩 精品 国产| 久久久国产成人免费| 免费在线观看影片大全网站| 国产精品美女特级片免费视频播放器 | 国产成人影院久久av| 久久久久网色| 中亚洲国语对白在线视频| 手机成人av网站| 欧美黑人欧美精品刺激| 脱女人内裤的视频| 国产免费现黄频在线看| 欧美人与性动交α欧美软件| 国产成人精品久久二区二区91| 色老头精品视频在线观看| 交换朋友夫妻互换小说| 黄网站色视频无遮挡免费观看| 大香蕉久久成人网| 女人被躁到高潮嗷嗷叫费观| 无人区码免费观看不卡 | 伦理电影免费视频| 男女床上黄色一级片免费看| 精品高清国产在线一区| 久久中文字幕一级| 激情在线观看视频在线高清 | 国产高清videossex| 久久精品国产亚洲av香蕉五月 | 黄网站色视频无遮挡免费观看| 免费人妻精品一区二区三区视频| 天堂俺去俺来也www色官网| 欧美乱妇无乱码| 999久久久国产精品视频| 三级毛片av免费| 韩国精品一区二区三区| 桃花免费在线播放| 在线观看一区二区三区激情| 免费久久久久久久精品成人欧美视频| 国产欧美日韩一区二区三| 久久热在线av| 国产精品九九99| 99久久99久久久精品蜜桃| 国产有黄有色有爽视频| 两性午夜刺激爽爽歪歪视频在线观看 | 精品国产一区二区三区四区第35| 法律面前人人平等表现在哪些方面| 日韩欧美一区二区三区在线观看 | 咕卡用的链子| 精品免费久久久久久久清纯 | 国产av一区二区精品久久| 久久午夜综合久久蜜桃| 久久久水蜜桃国产精品网| 国产精品电影一区二区三区 | 成人国产av品久久久| 熟女少妇亚洲综合色aaa.| 在线观看免费视频日本深夜| 一本—道久久a久久精品蜜桃钙片| 日韩制服丝袜自拍偷拍| 波多野结衣一区麻豆| 少妇 在线观看| 宅男免费午夜| 国产精品久久电影中文字幕 | 亚洲精品国产色婷婷电影| 久久狼人影院| 国产成人欧美在线观看 | 欧美老熟妇乱子伦牲交| 久久久国产一区二区| 国产精品久久久av美女十八| 大码成人一级视频| 一个人免费看片子| 女人精品久久久久毛片| 韩国精品一区二区三区| 在线观看免费午夜福利视频| 999久久久精品免费观看国产| 青草久久国产| 国产亚洲一区二区精品| 国产色视频综合| 天堂俺去俺来也www色官网| 最近最新免费中文字幕在线| 老司机在亚洲福利影院| 久久人妻福利社区极品人妻图片| 色综合婷婷激情| 亚洲熟妇熟女久久| 多毛熟女@视频| 狠狠精品人妻久久久久久综合| 一级片免费观看大全| 欧美性长视频在线观看| 国产一卡二卡三卡精品| 欧美人与性动交α欧美精品济南到| 午夜两性在线视频| 91国产中文字幕| 国产精品久久久久久精品古装| 一级片免费观看大全| 黄色视频在线播放观看不卡| 国产成人精品无人区| 国产老妇伦熟女老妇高清| 香蕉丝袜av| 亚洲自偷自拍图片 自拍| 亚洲人成77777在线视频| 搡老熟女国产l中国老女人| 岛国毛片在线播放| 亚洲色图综合在线观看| 成年人免费黄色播放视频| 99精品在免费线老司机午夜| 久久久久精品人妻al黑| 亚洲欧美日韩高清在线视频 | 中文字幕最新亚洲高清| 一进一出好大好爽视频| √禁漫天堂资源中文www| 精品国产一区二区三区四区第35| 十分钟在线观看高清视频www| 国产在线观看jvid| 国产男女内射视频| 午夜福利一区二区在线看| 色尼玛亚洲综合影院| h视频一区二区三区| 国产精品 欧美亚洲| 日韩 欧美 亚洲 中文字幕| 高清av免费在线| 在线亚洲精品国产二区图片欧美| 日本wwww免费看| 在线观看人妻少妇| 久久久久国产一级毛片高清牌| 99re在线观看精品视频| 国产高清videossex| 欧美午夜高清在线| 淫妇啪啪啪对白视频| 日本精品一区二区三区蜜桃| 啦啦啦视频在线资源免费观看| 日本黄色日本黄色录像| 亚洲av片天天在线观看| 99在线人妻在线中文字幕 | 9191精品国产免费久久| 丝袜美腿诱惑在线| 中文字幕高清在线视频| 精品免费久久久久久久清纯 | 亚洲国产中文字幕在线视频| 国产在线观看jvid| 男女高潮啪啪啪动态图| 国产欧美日韩一区二区三| 9色porny在线观看| tube8黄色片| 首页视频小说图片口味搜索| 少妇 在线观看| 波多野结衣一区麻豆| 欧美+亚洲+日韩+国产| 99久久国产精品久久久| 超碰成人久久| 中文字幕人妻丝袜制服| 不卡av一区二区三区| cao死你这个sao货| 又紧又爽又黄一区二区| 午夜精品久久久久久毛片777| 天天躁夜夜躁狠狠躁躁| 99精品在免费线老司机午夜| 不卡一级毛片| av网站免费在线观看视频| 在线观看66精品国产| 后天国语完整版免费观看| 妹子高潮喷水视频| 国产成人av教育| 国产一区有黄有色的免费视频| √禁漫天堂资源中文www| 日韩欧美一区二区三区在线观看 | aaaaa片日本免费| av在线播放免费不卡| 免费av中文字幕在线| 中文字幕人妻丝袜一区二区| 中文字幕高清在线视频| 免费观看人在逋| 一级a爱视频在线免费观看| 中文欧美无线码| 天天躁夜夜躁狠狠躁躁| 黄色片一级片一级黄色片| 老熟妇仑乱视频hdxx| 三级毛片av免费| 日本欧美视频一区| 亚洲欧洲日产国产| 热99久久久久精品小说推荐| 久久热在线av| 国产欧美日韩精品亚洲av| 国产精品电影一区二区三区 | 免费看a级黄色片| 亚洲av日韩在线播放| 午夜福利影视在线免费观看| 日本精品一区二区三区蜜桃| 久久99热这里只频精品6学生| 高清视频免费观看一区二区| 日本av手机在线免费观看| 女人久久www免费人成看片| 女人爽到高潮嗷嗷叫在线视频| 欧美精品人与动牲交sv欧美| 亚洲色图av天堂| 亚洲精华国产精华精| 国产亚洲av高清不卡| 青草久久国产| 老司机在亚洲福利影院| 欧美亚洲 丝袜 人妻 在线| av免费在线观看网站| 欧美日韩亚洲综合一区二区三区_| 国产精品国产高清国产av | 汤姆久久久久久久影院中文字幕| 亚洲精品在线美女| 青草久久国产| 成人国产一区最新在线观看| 日本wwww免费看| 免费黄频网站在线观看国产| 丰满人妻熟妇乱又伦精品不卡| 精品人妻熟女毛片av久久网站| 国产欧美日韩一区二区精品| 窝窝影院91人妻| 日韩三级视频一区二区三区| 免费在线观看日本一区| 成人亚洲精品一区在线观看| 18禁黄网站禁片午夜丰满| 99国产精品一区二区三区| 汤姆久久久久久久影院中文字幕| 12—13女人毛片做爰片一| 亚洲七黄色美女视频| 老司机亚洲免费影院| 精品国产一区二区三区久久久樱花| 热99国产精品久久久久久7| 一夜夜www| 变态另类成人亚洲欧美熟女 | 国产主播在线观看一区二区| 国产精品成人在线| 精品国内亚洲2022精品成人 |