• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A Review on Deep Learning Approaches to Image Classificationand Object Segmentation

    2019-08-13 05:54:40HaoWuQiLiuandXiaodongLiu
    Computers Materials&Continua 2019年8期

    Hao Wu,Qi Liu and Xiaodong Liu

    Abstract: Deep learning technology has brought great impetus to artificial intelligence,especially in the fields of image processing,pattern and object recognition in recent years.Present proposed artificial neural networks and optimization skills have effectively achieved large-scale deep learnt neural networks showing better performance with deeper depth and wider width of networks.With the efforts in the present deep learning approaches,factors,e.g.,network structures,training methods and training data sets are playing critical roles in improving the performance of networks.In this paper,deep learning models in recent years are summarized and compared with detailed discussion of several typical networks in the field of image classification,object detection and its segmentation.Most of the algorithms cited in this paper have been effectively recognized and utilized in the academia and industry.In addition to the innovation of deep learning algorithms and mechanisms,the construction of large-scale datasets and the development of corresponding tools in recent years have also been analyzed and depicted.

    Keywords: Deep learning,image classification,object detection,object segmentation,convolutional neural network.

    1 Introduction

    With the continuous improvement of computing system,such as the help of GPU and distributed computing system,the training of large-scale multi-layer artificial neural network has become possible.In addition,the project of ImageNet [Deng,Dong,Socher et al.(2009)] makes a fertile soil for deep learning to develop.It offers a very giant image database which contains 12 subtrees with 5247 synsets and 3.2 million images in total.This storage was going to bigger after it became public.More than a million of the images have a clear category tag and an annotation of the object in the image.The tags and annotations are up corrected or updated when mistakes found.The giant data quantity and highly tagging make the ImageNet almost become the standard the image processing in deep learning especially computer vision.Researchers from all the world compete their image classification,object detection and location algorithms in the ImageNet Large Scale Visual Recognition Competition (ILSVRC).The most attractive achievement at the beginning is the work from Krizhevsky et al.[Krizhevsky,Sutskever and Hinton (2012)]in 2012.They give a very efficient and effective network architecture that has a great impact on subsequent researches.This network was named as AlexNet by the researchers which is an homage to Yann LeCuns pioneering LeNet 5 network [Lecun,Boser,Denker et al.(1989)].They propose many efficient optimization techniques,such as Dropout,ReLU (Rectified Linear Unit) and LRN (Local Response Normalization).Many followup networks were put forward at the basement of the AlexNet such as VGG [Simonyan and Zisserman (2014)],GooLeNet [Szegedy,Liu,Jia et al.(2015)],ResNet [He,Zhang,Ren et al.(2016)].They extend depth and width of the AlexNet basically and optimize the structures of networks.These networks have a great impact on computer vision and pushed artificial intelligence technology to a new height.They are very inspiring for later researchers and have had far reaching influence.

    Deep learning has become a hot research topic for its excellent performance.More and more researchers from different fields devote their efforts into deep learning and combine it with their own research projects.Countless innovations have emerged constantly in recent years.For example,Google made a great progress with the neural machine in their translation systems [Wu,Schuster,Chen et al.(2016)].Further,Van Den Oord et al.[Van Den Oord,Dieleman,Zen et al.(2016)] design a WaveNet which can generate human likely voice from text.On the other hand,deep learning shows great strength in the field of games.The famous AlphaGo [Silver,Huang,Maddison et al.(2016);Silver,Schrittwieser,Simonyan et al.(2017)] defeat many human top Go masters in 2016 and 2017.Before that,the machine equipped with deep reinforcement learning technologies is realized the human like video games control in 2015.Even some robot algorithms have surpassed humans’ recordings [Mnih,Kavukcuoglu,Silver et al.(2015)].Besides,deep learning also shows its ability in artistic creation.GAN [Goodfellow,Pouget-Abadie,Mirza et al.(2014)] is a network that can generate fraudulent data based on training data.The generated images from the large data set are hard to find differences from the training data.Gatys et al.[Gatys,Ecker and Bethge (2015)] propose an artistic style transfer neural network.It can redraw the image according to other artistic works following the style of the learnt images.Besides,deep learning also shows its powerful performance in the tradition artificial intelligence such as NLP,auto driving,medical industry and robot industry et al.Innumerable specific neural networks and optimization methods have been invented.

    This paper mainly discusses the major progress of deep learning in image processing in recent years,especially in the fields of image classification,target detection and object segmentation.The three research areas have both connected and progressive relationships.The connected thing is that they are all based on the basic idea of a convolutional neural network.The progressive relationship is that their difficulty is getting higher and higher.Target detection and object segmentation all use some basic network models in image classification.The image classification algorithm based on convolutional neural network gives a lot of new ideas to target detection and object segmentation and achieves very good results.

    On the other hand,efficient experimental tools,high-performance computing platforms high-quality and large-scale training data sets are also an essential part of achieving an outstanding model.Therefore,this paper compares the typical experimental tools and training datasets in Section 2 firstly.To achieve better results,only increasing the width and depth of the network is far from enough.Next,this paper compares the typical deep learning networks developed in recent years and refine their advantages in Section 3.In the Section 4,the paper lists and compares several networks that have the most recognized results in the field of target detection.Finally,the Section 5 briefly describes the research progress in semantic segmentation and instance segmentation.

    2 Typical experimental tools and training datasets

    A workman must sharpen his tools if he is to do his work well.Good tools can make the research process go more efficiently and successful.For deep learning,the tools can be divided into two parts.The first part is a good computing platform,including hardware and software.Another part is the high-quality large-scale training datasets.

    2.1 Typical experimental tools

    Deep learning is a class of computationally intensive algorithms.They rely on high performance computing resources heavily,especially floating-point operations.Fortunately,with the development of technology,the floating-point operation ability of GPU has been emphasized in the industry.More and more researchers are using GPU and even distributed GPU clusters to speed up the training of large-scale artificial neural networks [Li,Zhang,Huang et al.(2016)].Therefore,to facilitate the adoption of the high-performance computing resources well and help researchers to focus on the implement of algorithms agilely,a variety of programming tools and frameworks emerge as the times require.Besides,most of these tools are not only for deep learning.They can also do a lot different kinds of scientific speed-up calculation in a very easy way.

    The frameworks in the Tab.1 are very popular both in the industry and academia.They provide rich convenient interfaces for mathematical computation.Users can use these interfaces to build their own neural network models conveniently with powerful GPU/CPU even their clusters without worrying about tedious computing details.

    Some of these frameworks lay particular emphasis on the underlying design of mathematical interfaces.Therefore,there is still a lot of work to do before users realize their own experiments.To further simplify programming,a more advanced framework arises at the historic moment.The high-level frameworks,such as Keras,use some of above frameworks as the backend so that combine ease of use and high performance together but sacrifice certain expansibility.

    Table1:Comparison of programming tools supporting deep learning

    Table2:Common datasets of image processing

    As the excellent performance of deep learning is more and more favored by the market,the customized chips and computing systems come into being,such as NVIDIA DGX with NVIDIA Tesla V100s,Google TPU.However,they are all expensive or no public for sale.In addition to speeding up on general-purpose chips,researchers are also trying to design specialized chips for accelerated calculations,such as FPGA,ASIC,and achieved good performance [Nurvitadhi,Sheffield,Sim et al.(2017)].What is more,researchers have already successfully ported models to embedded devices [Hegde,Ramasamy and Kapre (2016)] and mobile devices [Tsung,Tsai,Pai et al.(2016)].

    2.2 Typical data set comparison

    The development of deep learning cannot be separated from the development of data sets.The next table shows the typical dataset of image processing fields.They all play an important role in the recent neural network researches whatever in industry application or academic research.

    Tab.2 shows the common data sets of image processing which is used frequently.Some of these data sets have been unable to meet the needs of modern machine learning but they are still popular,such as MNIST.Because they can still verify the basic function of algorithms just like the first “hello world” program or be the standard of the contrasts in some performances of different algorithms.Some of the data sets were continually updated,such as ImageNet and PASCAL VOC.But the updates are stopped now.Because they cannot meet the needs like the MNIST.

    The Fig.1 shows some pictures of several typical image datasets.The picture (a) is some handwritten digits samples from MNIST,which is a subset of National Institute of Standards and Technology (NIST).The digits have been size-normalized and centered in a fixed-size image.This data set has a long history,and there are only ten kinds of pictures,but the quality of data set is very high.So many researchers regard it as the simplest data sets for testing algorithms.It is a very simple and meaningful benchmark in the field of image classification.The pictures (b) and (c) are two more modern practical datasets.They are very suitable for modern artificial neural network training.The ImageNet is a large-scale hierarchical image database which aims to populate the majority of the 80,000 synsets of WordNet with an average of 500-1000 clean and full resolution images.The hierarchical structure of each node is depicted by thousands of images.An average of more than five hundred of each node.They are all annotated by manual labelling.When ImageNet was born,deep learning based on big data has not been widely concerned.It's a very forward-looking dataset.On the basis of ImageNet,ImageNet Large Scale Visual Recognition Competition (ILSVRC) also has a great impact on the promotion of machine learning and artificial intelligence.Researchers and enterprises from all over the world compete against their algorithm performance on the basis of ImageNet.This competitive activity has greatly promoted the progress of related intelligent algorithms.At the same time,similar competitions emerge in an endless stream.Many datasets of other research fields are emerged accompanied by related competitions that attract a lot of researchers’ interests from all over the world.For example,The Microsoft COCO (Common Objects in Context) [Lin,Maire,Belongie et al.(2014)] dataset is designed not only for image classification,but also object recognition in the context of the broader question of scene understanding.Specifically,the COCO can be used to do object segmentation and recognition in context.Most pictures have multiple objects which are very different from ImageNet that only each object per image.It has more than 300,000 images and more than 2 million instances in total.The objects can be classified into 80 categories and you can find more than 100,1000 people from the dataset.Besides,each image has 5 captions in texts.This dataset aims at scene understanding,which is mainly intercepted from complex everyday scenes.The targets in the images are calibrated by precise segmentations.

    Figure1:Samples of some typical datasets

    Most of the deep learning algorithms are typical supervised machine learning.Therefore,accurate high-quality training samples are very important,which can be as important as the algorithm.Not a single one can be omitted.Collecting so much data,classifying and tagging it one by one is a quite great project.Thanks to the authors of these open source data sets for their unselfish dedication to their work which is meaningful.These data sets have greatly promoted the development of artificial intelligence technology.Many researchers,engineers,and students with limited funds and strength can easily verify their innovation in deep learning with these open source data set and made a leap forward.Since then,CNN has become a hot research topic in the academic world.The COCO dataset has 91 categories,although there are fewer than the ImageNet category,but there are more images in each category,which is conducive to obtaining more capabilities in each category in a particular scene.Compared with PASCAL VOC,there are more categories and images.

    The researchers from Google and CMU [Li,Wang,Agustsson et al.(2017)] find that the quality and quantity of data are crucial to the training of deep learning models.This conclusion is based on an already existing JFT image dataset [Sun,Shrivastava,Singh et al.(2017)].The data set was first proposed by scientists such as Geoffrey Hinton et al.and expanded by Hinton et al.[Hinton,Vinyals and Dean (2015)].The JFT dataset has over 300 million images and is marked with 18291 categories.It is much larger than the size of ImageNet,but the accuracy is not high.They find that the performance of visual tasks continues to increase linearly with the magnitude of the training data size.Network performance increases linearly with the magnitude of training data.Representation learning (or pre-training) is still of great use.By training a better basic model,it can improve the performance of visual tasks.Besides the hardware performance of computer system,the training datasets are also very important to deep neural network models.

    3 An analysis of convolutional neural network based image classification

    Image classification means that the image is structured into a certain category of information,and the image is described with a previously determined category or instance ID.This task is the simplest and most basic image understanding task,and it is also the task of the deep learning model to achieve the first breakthrough and realize large-scale application.Among them,ImageNet is the most authoritative evaluation set.Each year,ImageNet Large Scale Visual Recognition Competition (ILSVRC) has spawned a large number of excellent indepth network structures,providing the basis for other tasks.In the application field,the recognition of faces,scenes,etc.can be classified as classification tasks.

    In the domain of image classification,the convolutional neural network shows excellent performance.In general,the CNNs are the state-of-the-art contrast to the classic algorithms.The structural features of the convolutional neural network are more suitable for solving the problem of image field.Through the continuous research and improvement of its structure,a series of network models have been formed,which have been successful in a wide range of practical applications.The convolutional neural network can reduce the parameters needed to learn by using the spatial structure relation,thus improving the training efficiency of the back-propagation algorithm.In the convolutional neural network,the first volume layer will accept a pixel level image input,each operation only a small image processing convolution,and convolution change and then spread to the back of the network,each layer of convolution will have the most effective feature extraction data.This method can extract feature the most basic image,such as edges or corners in different directions,and then combined and abstract formation characteristics of higher order,so CNNs can cope with various situations,the theory of image with zoom,rotation and translation invariance.

    LeCun has used the back-propagation algorithm to train multi-layer neural networks to identify handwritten postcodes,which is the first modern CNN [Lecun,Bottou,Bengio et al.(1998)] in 1998.However,limited to the development of computing performance and data sets,it was not until 2012 that Krizhevsky et al.[Krizhevsky,Sutskever and Hinton(2012)] proposed a CNN to adapt to the large data set.On the millions of ImageNet data sets,the effect is much more than the traditional method,and the classification accuracy has risen from more than 70% to more than 80%.

    Table3:Architectures of typical convolutional neural networks

    After that,because of the great performance,CNNs have attracted the attention of many researchers.Until the last ILSVRC in 2017,the image classification accuracy of the deep learning algorithm on ImageNet has approached or even surpassed that of humans.

    Because of this,the ILSVRC competition is no longer continuing.Researchers will focus on more challenging projects,such as WebVision Challege [Ioffe and Szegedy (2015)].Tab.3 shows the best CNNs in the ILSVRC competition from 2012.It lists the new features of the successive generations of networks.As we can see from the table,more and more optimizations are applied into the network design,such as ReLU,Dropout,Inception,Local Response Normalization (LRN),Batch Normalization,etc.With the development of the network,not all of the optimization methods are applied to the latest algorithms.For example,not all the networks use the LRN method.The author of VGG believes that LRN does not play an optimization role in their network structure.However,the Inceptions [Szegedy,Vanhoucke,Ioffe et al.(2016);Szegedy,Ioffe,Vanhoucke et al.(2017);Girshick (2014)] all adopt the LRN.On the other hand,the Dropout method has always been considered to be able to effectively improve the network generalization ability and reduce overfitting.But after the first version of the Inception,the method has been abandoned.Because batch normalization in the Inceptions regularizes the model and reduces the need for Dropout.It can be either removed or reduced in strength in a batchnormalized network

    The performance results of the above networks are based on the ImageNet data set in the ImageNet Large Scale Visual Recognition Competition (ILSVRC).However,the competition stopped after the 8 sessions from 2009 to 2017.The accuracy rate of identifying objects from the original algorithm is only 71.8% up to 97.3% of the present,and the error rate of recognition is far below the 5.1% of human.Although the ImageNet challenge has ended its short life cycle,ImageNet data set will continue to exist.Up to now,there are more than 13 million pictures and will grow in the future and continue to contribute to the field of computer vision.In the future,ImageNet will remain open for free use by researchers.Even if the ImageNet competition itself is over,its legacy will continue to affect the entire industry.Since 2009,dozens of newly developed datasets have introduced computer vision,neural language processing and speech recognition and other subdomains.ImageNet has changed researchers’ thinking mode.Although many people still care about models,they are also concerned about data.Data redefine our way of thinking about models.

    4 Two significant target detection ideas with CNNs

    It is not enough to just classify images.Classification is the basic of computer vision.Object localization,object recognition,semantic segmentation,instance segmentation and key point detection are more hard and meaningful tasks.The classification task is concerned with the whole,given the content description of the entire picture,while the detection is focused on the specific object target,and it is required to obtain the category information and location information of this target at the same time.Compared with classification,target detection gives the understanding of the foreground and background of the image.The algorithm needs to separate the target of interest from the background and determine the description of the target,such as the category and location of the target.Therefore,the output of the target detection model is a list.Each item of the list uses a data group to give the category and position of the target,which is commonly represented by the coordinates of the rectangular detection frame.

    Target detection research has been conducted for many years,and there are many methods that have been widely recognized and applied in the industry,such as Jones et al.[Jones and Viola (2001);Zhu,Yeh,Cheng et al.(2006)].But most classic methods are very dependent on finding effective features which is hard to be found in a common simple method.The introduction of CNNs changed the main research ideas in the field of target detection.It frees researchers from complex feature engineering.In the next part,several typical target detection algorithms are introduced in the field of target detection with the help of deep learning.

    At present,there are two main ideas in the field of target detection based on deep learning.One is the RCNN series method based on the region proposal,and the other is the YOLO series algorithm based on the regression.

    4.1 RCNNs:regions with CNN features

    4.1.1 RCNNs:Regions with CNN features

    The classical target detection algorithm uses a sliding window method to determine all possible areas in turn.Girshick [Girshick (2015)] extract a series of more likely candidate regions in advance with the method of selective search,and then the CNNs based extraction features are used only in these candidate regions for judgment.The RCNN algorithm can be divided into 4 steps:

    1.Candidate region generation:An image generates 1K~2K candidate regions (using the Selective Search method).

    2.Feature extraction:Using deep convolution network to extract feature (CNN) for each candidate region.

    3.Category judgment:The feature is sent to each class of SVM classifier to distinguish whether it belongs to the class.

    4.Location:Refinement using regressor fine correction candidate frame position.

    The drawback of this method is that RCNN has the problem of repeated computation.There are thousands of regions in proposal,most of which overlap each other,and the overlapped parts will be repeatedly extracted from feature.To solve the problem,Kaiming He et al.[He,Zhang,Ren et al.(2014)] propose the SPP-Net to improve the RCNN.

    4.1.2 SPP-net:spatial pyramid pooling in deep convolutional networks.

    An image has 1~2 k candidate boxes in RCNN,and each one has to enter a CNN to do convolution.SPP-net proposes to extract the RoI (Region of Interest) features on the feature map,so that only a convolution is needed on the entire image.The overall process is similar to RCNN.But because all Z the features of the RoI are extracted directly from the feature map,the convolution operation is greatly reduced,and the efficiency is improved.

    Figure2:SPP-net data flow graph

    4.1.3 FAST and FASTER RCNN

    Girshick et al.combined with SPP-net to improve the performance of RCNN networks[Girshick (2015);Ren,He,Girshick et al.(2015)].The Faster RCNN even realize object recognition in video.Specifically,in Fast-RCNN,the author put bounding box regression into the internal neural network and classified it into region and became a multi-task model.The actual experiment also proved that these two tasks can share convolutional feature and promote each other.However,the performance can still be promoted.Faster-RCNN is an end-to-end CNN object detection model.The authors propose that the convolutional level features in network can be used to predict category dependent region proposal,and do not need to perform algorithms such as selective search in advance.The author integrates region proposal extraction and Fast-RCNN part into a network model.Although the training stage is still multi-step,the detection phase is very convenient and fast,and the accuracy rate is not much different from the original Fast-RCNN.At last,for the very deep VGG-16 model,the Faster-RCNN detection system has a frame rate of 5fps (including all steps) on a GPU in 2016.

    4.1.4 Mask R-CNN

    Mask-RCNN [He,Gkioxari,Dollar et al.(2017)] is a parallel detection and segmentation results,which means that the two results can be got at one time,but unlike the previous segmentation after do classification.The general framework of Mask-RCNN is still like the Faster-RCNN framework.It can be said that the fully connected subdivision network is joint after the basic feature network.The task of the network is from the original two tasks (classification+regression) to three tasks (classification+regression+segmentation).

    Figure3:The Mask R-CNN framework for instance segmentation

    Mask R-CNN is mainly divided into two stages:

    Stage 1:Generate the candidate box area.This process is the same as Faster R-CNN,the RPN (Region Proposal Network) is used.

    Stage 2:RoIPool is used in the candidate box area to extract features and to classify and border box regression,and a two-element mask is generated for each RoI.

    In target detection,there are some errors in the given bounding box and the original graph,which does not affect the results of the classification detection.But in pixel level image segmentation,such spatial location error will seriously affect the segmentation results.Therefore,this network proposes the use of bilinear interpolation to solve this problem,that is,RoIAlign.The image is passed through RoIAlign instead of RoIPool,making the area of the feature map selected by RoIPool more accurately corresponding to the area of the original image.

    Mask R-CNN is simple to train and adds only a small overhead to Faster R-CNN,running at 5 fps.Moreover,Mask R-CNN is easy to generalize to other tasks,e.g.,allowing us to estimate human poses in the same framework.

    Although the detection accuracy of RCNNs is high,if it is applied to video information processing,the excessive calculation amount required by the algorithm makes it difficult to process the video information in real time in a single machine condition.The emergence of YOLOs has brought the target detection speed to a new level with the lowest loss of accuracy rate.

    4.2 YOLO:you only look once

    YOLO [Redmon,Divvala,Girshick et al.(2015)] is a convolutional neural network that can predict multiple Box locations and categories at a time.It can achieve end to end target detection and recognition,and its biggest advantage is fast speed.In fact,the essence of target detection is regression,so a CNN that implements a regression function does not require a complex design process.YOLO does not train the network in the way of selecting sliding windows or extracting proposal,but directly selecting the whole training model.The advantage of this way is that it can better distinguish the target and background area.In contrast,the Fast R-CNN trained by proposal often mistakenly checks the background area as a specific target.Of course,YOLO has sacrificed some precision while improving the speed of detection.Until now,the YOLO series of algorithms have developed three generations

    4.2.1 YOLO Version 1:beginning of a regression-based target detection algorithm

    YOLO [Redmon,Divvala,Girshick et al.(2015)] is the most popular object detection algorithm.It is fast and simple.As the name implies,this algorithm recognizes all objects by looking at the image only once.This algorithm can achieve real-time object detection,about 40 frames per second with the help of the Titan X GPU.The accelerated version of YOLO is almost 150 fps.YOLO reasons globally about the image when making predictions.It can also learn generalizable representations of objects which mean that its generalization ability is relatively strong.YOLO supports end-to-end training,which will reduce a lot of unnecessary work,and the entire model is actually a convolutional neural network.

    YOLO’s computing process can be divided into the following steps:

    Step1:Divide the original image into an S×S grid.

    Step2:Each grid predicts B bounding boxes and confidence scores which can be represented by,whererepresents the probability that the current position is an Object,IOU is the overlap probability between the predicted box and ground truth.The x,y is the center coordinate and the w,h is the size of the box.

    Step3:The probability of each grid prediction class.

    Step4:When predicting,multiply the class conditional probability and confidence:

    Similar to R-CNN and DPM,when a large object is encountered,it is still necessary to perform non-maximally suppressed operations.Since B has a value of 2,that is,a grid will only return two boxes,and a grid will have only one category,so if there are multiple categories in a grid,there will be problems.YOLO has better results for images like small objects such as a flock of birds.

    Figure4:The YOLO models detection as a regression problem.It divides the image into an S×S grid and for each grid cell predicts B bounding boxes,confidence for those boxes,and C class probabilities.These predictions are encoded as an S×S×(B?5+C) tensor

    4.2.2 YOLO Version 2:a more comprehensive improvement

    There are still many places in the original YOLO that can be improved,so the original author made many improvements in the first version and achieved remarkable results.In summary,there are two major improvements.Because the network can distinguish approximately 9000 objects under combining training with ImageNet and COCO,it is also called YOLO9000 [Redmon and Farhadi (2016)].

    First,the author used a series of methods to improve the original YOLO multi-objective inspection framework.With the advantage of maintaining the original speed,the accuracy was improved.Using the VOC 2007 data set test,the mAP (Mean Average Precision) at 67 FPS reached 76.8%,and the mAP at 40 FPS reached 78.6%.The overall performance is comparable to Faster R-CNN and SSD.

    Second,the author proposes a joint training method for target classification and detection.Through this method,YOLO9000 can simultaneously train in COCO and ImageNet.The trained model can achieve real-time detection of up to 9000 objects.

    Some specific improvements are as follows:YOLOv2 adds batch normalization behind the volume base layer,removes the dropout layer,and increases its mAP by 2%.During network training,the network is changed from 224*224 to 448*448,and in order to ensure that there are only an odd number of positioning positions in the feature map,only one central cell is guaranteed,and the network is finally set to 416*416.Finally achieved 4% mAP increase.In addition,previous YOLO uses the data of the full connection layer to complete the prediction of the border,resulting in the loss of more spatial information and inaccurate positioning.In this version,the author draws on the anchor idea of Faster R-CNN,removes the full connectivity layer from the network,and combines Dimension Clusters and Direct location prediction to improve the mAP by 5%.

    In terms of training,unlike the method of fixing the picture size of the input network,the author finetunes the network after several iterations.Every 10 epochs,new image sizes are randomly selected.The down-sampling parameter used by the YOLO network is 32,then a scaled multiple of 32 is used for {320,352...608}.The final minimum size is 320*320 and the largest size is 608*608.Then adjust the network according to the input size for training.This mechanism allows the network to better predict pictures of different sizes,meaning that the same network can perform detection tasks with different resolutions.YOLOv2 runs faster on small pictures,achieving a balance in speed and accuracy.

    4.2.3 YOLO Version 3:the latest improvement

    The YOLOv3 model is much more complex than the previous model and can be weighed against the speed and accuracy by changing the size of the model structure.It improved multi-scale prediction,and better basic classification networks and classifiers.YOLOv3 does not use Softmax to classify each box.Instead,it is replaced by multiple independent logistic classifiers,and the accuracy does not decrease.

    4.3 General comparison of RCNNs and YOLOs

    Compared to the RCNNs methods,the YOLO series methods have the following advantages:

    · High speed as previous introduction.

    · False positive rate of background is low.

    · Versatility.YOLO also applies to object detection in artistic works.Its detection rate for non-natural image objects is much higher than DPM and RCNN series detection methods.

    Tab.4 lists the main features of RCNNs and YOLOs.In general,from R-CNN,SPP-NET,Fast R-CNN to Faster R-CNN,and from YOLOv1 to YOLOv2,the detection process based on deep learning has become increasingly streamlined and accurate.The speed is getting faster and faster.It can be said that the R-CNN series target detection methods based on the region proposal and the YOLO series target detection methods based on the regression are the two main branches in the current target detection field.

    The target detection,such as RCNNs and YOLOs,only gives the positions and labels of the objects in the image.However,many times,there is a need to detect the edges of objects and give relevant descriptions of the objects at the same time.The re-search of object segmentation is to achieve this goal.

    Table4:Features of RCNNs and YOLOs

    5 Pixel-level object segmentation

    Object detection only gives the locations and labels of the objects,which is not specific enough.It is more difficult to separate out all the pixels related to the object and give the categories.This operation is called object segmentation.Object segmentation includes semantic segmentation and instance segmentation.The former is an extension of the prebackground segmentation.It requires the separation of image parts with different semantics,while the latter is an extension of the detection task and requires the outline of the objects,which is more refined than the detection frame.Object segmentation is a pixel-level description of an image.It gives each pixel category meaning and is suitable for understanding demanding scenes,such as the segmentation of roads and non-roads in auto pilot,geographic information system,and medical image analysis,etc.

    5.1 Semantic segmentation

    Before deep learning was developed,the semantic segmentation method was diverse,and the effect levels were uneven,such as thresholding methods,clustering-based segmentation methods,graph partitioning segmentation methods and even the pixel-level decision tree classification [Shotton,Johnson and Cipolla (2008);Shotton,Fitzgibbon,Cook et al.(2011);Shi and Malik (1997);Rother,Kolmogorov and Blake (2004)].After computer vision entered the era of deep learning,semantic segmentation also entered a new stage of development.A series of semantic segmentation networks based on convolutional neural networks represented by Fully Convolutional Networks (FCNs) are proposed and repeatedly refresh the semantic segmentation accuracy of images.

    5.1.1 Full convolutional neural network (FCN) for semantic segmentation

    The idea of the FCN [Long,Shelhamer and Darrell (2015)] is very intuitive.It directly performs pixel-level end-to-end semantic segmentation.It can be implemented based on the mainstream deep convolutional neural network model.It reuses ImageNet’s pretraining network for se-mantic segmentation and uses deconvolutional layer up-sampling.At the same time,it also introduced a skip connection to improve the up-sampling of coarse pixel positioning.

    The SegNet [Badrinarayanan,Kendall and Cipolla (2017)] improves use of encoder and decoder.It applies the result of the pooling layer to the decoding process.With other improvements,the segmentation accuracy is slightly better than that of the FCN,and the overall efficiency is slightly higher than that of the FCN.

    5.1.2 Dilated convolutions

    One deficiency of the FCN is that due to the presence of the pooling layer,the size of the tensor,i.e.,the length and width,becomes smaller and smaller.However,the original design of the FCN required an output that was consistent with the input size,so the FCN did an upsampling.But up-sampling cannot retrieve lost information completely without loss.

    Dilated Convolution [Yu and Koltun (2015)] is a good solution for this.Since pooled down-sampling operations can result in information loss,the pooling layer is removed directly.However,the removal of the pooling layer will bring about a smaller receptive field in each layer of the network,which will reduce the prediction accuracy of the entire model.The main contribution of Dilated Convolution is to remove the down-sampling operation of the pool without reducing the receptive field of the network.

    5.1.3 Following-up development

    After this,most semantic separation networks are inseparable from the FCN architecture.The DeepLab [Chen,Papandreou,Kokkinos et al.(2018);Chen,Papandreou,Schroff et al.(2017)] also uses Dilated Convolutions and fully connected CRF added and proposes atrous spatial pyramid pooling.Dilated Convolution has several disadvantages,such as large amount of computation and a large amount of memory.Therefore,RefineNet [Lin,Milan,Shen et al.(2017)] designs the Encoder-Decoder architecture with well thoughtout decoder blocks and all the components follow residual connection design which reduces the computing requirements.Besides,in the recent years,more and more improvements are emerged,such as PSPNet [Zhao,Shi,Qi et al.(2017)],Large Kernel Matters [Peng,Zhang,Yu et al.(2017)] and DeepLab v3 [Chen,Papandreou,Schroff et al.(2017)].The semantic segmentation is developing fast.However,there are still many challenges waiting to be resolved.

    Semantic segmentation technology based on deep learning can achieve a segmentation effect that is faster than traditional methods,but its requirement for data annotation is too high.Not only does it require huge amounts of image data,they also need to provide Semantic labels that are accurate to the pixel level.Therefore,more and more researchers have begun to turn their attention to the problem of image semantic segmentation under Weakly-supervised conditions.In this kind of problem,the image only needs to provide the image level annotation,and the semantic segmentation accuracy comparable to the existing method can be obtained without the need of expensive pixel level information.

    5.2 Instance segmentation

    Instance segmentation is a complex of object detection and semantic segmentation.It can both detect the object,give its bounding box,and it can be segmented to the edge of the object.Relative semantic segmentation,instance segmentation can label different individuals of the same type of object on the picture.Therefore,instance segmentation is a very comprehensive problem that combines object detection,semantic segmentation and image classification.

    The MASK R-CNN and FCIS [Li,Qi,Dai et al.(2017)] are the most significant research results in the past two years.The MASK R-CNN has been introduced in section 4.1.In fact,this method can effectively detect the simultaneous occurrence of each target and generate a high-quality segmentation mask for each instance.Mask R-CNN has a simple and straightforward idea:For Faster R-CNN,it has two outputs for each target object.One is the class label and the other is the bounding-box offset.Based on this,the Mask R-CNN method adds the output of the third branch:the object mask.The difference between the object mask and the existing class and box output is that it requires a finer refinement of the spatial layout of the object.

    FCIS solves the problem of instance segmentation through a multitasking network.It inherits all the merits of FCNs for semantic segmentation [Dai,He,Li et al.(2016)] and instance mask proposal First,primary features are first extracted through a convolutional neural network,and a RoI (Region of Interest) is proposed.Then,for each RoI region,corresponding features are extracted by RoI Warping and RoI Pooling.Next,use the full connected layer to perform the foreground and background division.Finally,use the full connected layer for image classification for each RoI.

    Figure5:FCIS vs. Mask R-CNN.FCIS exhibits systematic artifacts on overlapping objects,suggesting that it is challenged by the fundamental difficulty of instance segmentation.Mask R-CNN shows no such artifacts

    There are three shortcomings of the instance-based partitioning architecture based on proposal.First,if two objects share the same or similar boundaries,the algorithm cannot accurately identify them,especially for low-fill-rate linear objects such as windmills and windows.Second,there is no solution in the architecture that can prevent two instances from sharing pixels.Finally,the number of identifying instances is usually limited by the number of proposals that the network can handle.Some researchers took other ideas to avoid the above problems,such as the idea of instance embedding.Each pixel in the network output is a point in the embedding space.Points belonging to the same object are relatively close in the embedding space,while points belonging to different classes are far apart in the embedding space.Each pixel in the network output is a point in the embedded space [De Brebandere,Neven and Van Gool (2017);Fathi,Wojna,Rathod et al.(2017);Kong and Fowlkes (2017)].The drawback is that compared to the methods based the idea of proposal,the results of these methods are not as good as Mask R-CNN or FCIS at present.

    6 Conclusion

    This paper has briefly reviewed typical deep convolutional neural networks in recent years and compared their differences and similarities.Effective network structures and optimization methods in these convolutions neural network is summarized.Besides,target detection and object segmentation algorithms based on deep convolution neural network are also summarized.These models have been becoming or will be recognized as new hotspots in deep learning and convolutional neural networks to effectively solve problems in computer vision,multi-object classification and/or relevant fields in these years and are therefore recognized as effective methods and/or de-facto tools in the industry and academia.

    Besides the performance of the algorithm,however,most algorithms still highly rely on their own training datasets,which directly determines the functionality and performance of the algorithms.In that case,careful preparation of a dataset becomes critical and even tricky sometimes,so that it is imminent to develop migration learning methods based on deep learning.

    Acknowledgment:This work has received funding from 5150 Spring Specialists(05492018012),the European Union Horizon 2020 research and innovation programme under the Marie Sklodowska Curie grant agreement no.701697,Major Program of the National Social Science Fund of China (Grant No.17ZDA092),the PAPD fund,and 333 High-Level Talent Cultivation Project of Jiangsu Province (BRA2018332).

    简卡轻食公司| 你懂的网址亚洲精品在线观看 | 亚洲丝袜综合中文字幕| 久久精品91蜜桃| 久久久久久久久中文| 人人妻人人看人人澡| 久久亚洲国产成人精品v| 国产精品野战在线观看| 欧美高清成人免费视频www| 国产又色又爽无遮挡免| 欧美潮喷喷水| 老司机福利观看| 男插女下体视频免费在线播放| 美女内射精品一级片tv| 男女视频在线观看网站免费| 午夜免费激情av| 午夜精品国产一区二区电影 | 日本av手机在线免费观看| 最近2019中文字幕mv第一页| 日韩在线高清观看一区二区三区| 欧美精品一区二区大全| 丰满人妻一区二区三区视频av| 少妇猛男粗大的猛烈进出视频 | 久久久久网色| 成人欧美大片| 日产精品乱码卡一卡2卡三| 国产一区亚洲一区在线观看| 国产精品久久久久久精品电影小说 | 日韩一本色道免费dvd| 免费电影在线观看免费观看| 2021少妇久久久久久久久久久| 中文字幕人妻熟人妻熟丝袜美| 搞女人的毛片| 亚洲乱码一区二区免费版| 亚洲最大成人av| 国产麻豆成人av免费视频| 99热这里只有是精品50| 亚洲精品aⅴ在线观看| av又黄又爽大尺度在线免费看 | 欧美激情国产日韩精品一区| av播播在线观看一区| av在线天堂中文字幕| 热99re8久久精品国产| 日韩视频在线欧美| 两个人的视频大全免费| АⅤ资源中文在线天堂| 国产乱来视频区| 91精品一卡2卡3卡4卡| 2022亚洲国产成人精品| 91在线精品国自产拍蜜月| 蜜桃亚洲精品一区二区三区| 免费无遮挡裸体视频| www日本黄色视频网| 综合色丁香网| 99久国产av精品| 日韩,欧美,国产一区二区三区 | 级片在线观看| 亚洲国产高清在线一区二区三| 亚洲av.av天堂| 亚洲欧美清纯卡通| 国产成人a区在线观看| 人体艺术视频欧美日本| 美女黄网站色视频| 国产美女午夜福利| 国产成人freesex在线| 91aial.com中文字幕在线观看| 免费观看人在逋| 国产一级毛片在线| 黄色日韩在线| 日韩大片免费观看网站 | 熟妇人妻久久中文字幕3abv| 亚洲乱码一区二区免费版| 在线观看美女被高潮喷水网站| 女人被狂操c到高潮| 欧美激情国产日韩精品一区| 国产精品美女特级片免费视频播放器| 日本一本二区三区精品| 亚洲av日韩在线播放| 久久人人爽人人爽人人片va| 精品无人区乱码1区二区| 97超碰精品成人国产| 国产亚洲5aaaaa淫片| 听说在线观看完整版免费高清| 精品无人区乱码1区二区| 天堂av国产一区二区熟女人妻| 国产一级毛片七仙女欲春2| 免费观看a级毛片全部| 视频中文字幕在线观看| 亚洲国产精品久久男人天堂| 精品国产一区二区三区久久久樱花 | 国产精品久久久久久久久免| 少妇高潮的动态图| 国产伦精品一区二区三区视频9| 一区二区三区免费毛片| 欧美潮喷喷水| 日韩一本色道免费dvd| 国产视频首页在线观看| 成年免费大片在线观看| 人人妻人人看人人澡| 欧美日本亚洲视频在线播放| 狠狠狠狠99中文字幕| 蜜桃久久精品国产亚洲av| 丝袜喷水一区| 国产三级中文精品| 又粗又硬又长又爽又黄的视频| 亚洲欧美清纯卡通| 夜夜爽夜夜爽视频| 纵有疾风起免费观看全集完整版 | 久久精品国产自在天天线| 国产精品久久久久久av不卡| 在线播放国产精品三级| 欧美一区二区精品小视频在线| 国产一区亚洲一区在线观看| 一区二区三区免费毛片| 人妻系列 视频| 国产久久久一区二区三区| 老师上课跳d突然被开到最大视频| 别揉我奶头 嗯啊视频| 午夜激情福利司机影院| 直男gayav资源| 国产成人午夜福利电影在线观看| 女人久久www免费人成看片 | 天天躁日日操中文字幕| 午夜激情福利司机影院| 特大巨黑吊av在线直播| 小蜜桃在线观看免费完整版高清| 一级黄色大片毛片| 亚洲欧美精品自产自拍| 精品久久久久久久久亚洲| 男女边吃奶边做爰视频| 色播亚洲综合网| 建设人人有责人人尽责人人享有的 | 少妇人妻精品综合一区二区| 久久久久久久久久久免费av| 久久人人爽人人爽人人片va| 亚洲欧美精品综合久久99| 欧美+日韩+精品| 国产午夜精品久久久久久一区二区三区| 日日干狠狠操夜夜爽| 尤物成人国产欧美一区二区三区| 91午夜精品亚洲一区二区三区| 久久综合国产亚洲精品| 免费黄网站久久成人精品| av福利片在线观看| 亚洲欧美成人综合另类久久久 | 在线天堂最新版资源| 久久精品影院6| 成人av在线播放网站| 99热这里只有是精品50| 亚洲成av人片在线播放无| 国产真实伦视频高清在线观看| 性插视频无遮挡在线免费观看| 最近最新中文字幕大全电影3| 国内少妇人妻偷人精品xxx网站| 男女边吃奶边做爰视频| 亚洲av电影不卡..在线观看| 在线观看66精品国产| 变态另类丝袜制服| 中文字幕制服av| 夫妻性生交免费视频一级片| 天堂√8在线中文| 最近手机中文字幕大全| 欧美日本视频| 欧美日韩精品成人综合77777| 国产精华一区二区三区| 在线播放无遮挡| 草草在线视频免费看| 禁无遮挡网站| 国产在视频线在精品| 国产爱豆传媒在线观看| 免费观看性生交大片5| 精品人妻一区二区三区麻豆| 中文欧美无线码| 国产精品爽爽va在线观看网站| 麻豆乱淫一区二区| 亚洲成人av在线免费| 成年女人永久免费观看视频| 亚洲欧美一区二区三区国产| 久久久久国产网址| 国产精品日韩av在线免费观看| 春色校园在线视频观看| kizo精华| av又黄又爽大尺度在线免费看 | 不卡视频在线观看欧美| 中文精品一卡2卡3卡4更新| or卡值多少钱| 丝袜喷水一区| 只有这里有精品99| 搡老妇女老女人老熟妇| 秋霞伦理黄片| 欧美又色又爽又黄视频| 97热精品久久久久久| 欧美一级a爱片免费观看看| 亚洲欧美精品专区久久| 国产精品嫩草影院av在线观看| 成人美女网站在线观看视频| 国产一区二区三区av在线| 亚洲欧美清纯卡通| 中文字幕人妻熟人妻熟丝袜美| 国内精品宾馆在线| 看免费成人av毛片| 精品久久国产蜜桃| 三级国产精品片| 美女脱内裤让男人舔精品视频| 国产av不卡久久| 又爽又黄无遮挡网站| 国产精品久久久久久久电影| 国产视频首页在线观看| 免费观看的影片在线观看| 欧美潮喷喷水| 欧美一区二区精品小视频在线| 一夜夜www| 亚洲国产精品国产精品| 91aial.com中文字幕在线观看| 亚洲综合精品二区| 国产成人一区二区在线| 亚洲av福利一区| 国产中年淑女户外野战色| 夫妻性生交免费视频一级片| 亚洲av一区综合| av在线老鸭窝| av在线观看视频网站免费| 日日摸夜夜添夜夜爱| 久久精品国产亚洲av天美| 国产免费一级a男人的天堂| av在线播放精品| 99热精品在线国产| 男女下面进入的视频免费午夜| 亚洲精品影视一区二区三区av| 夜夜爽夜夜爽视频| 亚洲成人精品中文字幕电影| 亚洲最大成人手机在线| 三级男女做爰猛烈吃奶摸视频| 变态另类丝袜制服| 精品99又大又爽又粗少妇毛片| 久久久精品大字幕| 美女大奶头视频| 精品一区二区三区视频在线| 少妇高潮的动态图| 男女边吃奶边做爰视频| 久久久精品大字幕| 亚洲av男天堂| 欧美丝袜亚洲另类| 亚洲经典国产精华液单| 淫秽高清视频在线观看| 国产色婷婷99| 久久99热这里只有精品18| 日日摸夜夜添夜夜爱| 国产免费又黄又爽又色| 又粗又爽又猛毛片免费看| 久久午夜福利片| 国产伦理片在线播放av一区| 丰满人妻一区二区三区视频av| 91久久精品国产一区二区成人| 在线观看美女被高潮喷水网站| 久久99精品国语久久久| 久久6这里有精品| 99热这里只有精品一区| 国产精品.久久久| 亚洲综合色惰| 午夜福利视频1000在线观看| 国产成人精品一,二区| 久久热精品热| 亚洲欧洲日产国产| 成人av在线播放网站| 亚洲三级黄色毛片| 亚洲激情五月婷婷啪啪| 国产精品一区二区三区四区久久| 国产淫片久久久久久久久| a级一级毛片免费在线观看| 久久久亚洲精品成人影院| 国产精品三级大全| 午夜激情福利司机影院| 18+在线观看网站| 中文欧美无线码| 一级毛片我不卡| 天天一区二区日本电影三级| 中文字幕熟女人妻在线| 丝袜喷水一区| 国产一区二区在线观看日韩| 亚洲最大成人手机在线| 岛国毛片在线播放| 综合色丁香网| 又黄又爽又刺激的免费视频.| 看非洲黑人一级黄片| 一级黄片播放器| 国产免费福利视频在线观看| videossex国产| 中文资源天堂在线| 精品久久国产蜜桃| 欧美高清性xxxxhd video| 亚洲欧美精品专区久久| 春色校园在线视频观看| 国产熟女欧美一区二区| 亚洲最大成人中文| 国产精品一区二区三区四区久久| 中文字幕人妻熟人妻熟丝袜美| 婷婷色综合大香蕉| 三级毛片av免费| 国产老妇女一区| 丝袜美腿在线中文| 国产精品.久久久| 国产男人的电影天堂91| 中文字幕人妻熟人妻熟丝袜美| 日本与韩国留学比较| 水蜜桃什么品种好| 天堂中文最新版在线下载 | 国产探花极品一区二区| 国产精品爽爽va在线观看网站| 一级爰片在线观看| 99久久精品热视频| 18禁动态无遮挡网站| av播播在线观看一区| 国产片特级美女逼逼视频| 在线天堂最新版资源| 日本av手机在线免费观看| 91精品一卡2卡3卡4卡| 在线免费观看的www视频| a级毛色黄片| 国产三级中文精品| 亚洲四区av| 99视频精品全部免费 在线| 免费看美女性在线毛片视频| 国产又色又爽无遮挡免| 99在线人妻在线中文字幕| 麻豆精品久久久久久蜜桃| 久久精品91蜜桃| 欧美激情在线99| 两性午夜刺激爽爽歪歪视频在线观看| 最近视频中文字幕2019在线8| 久久这里有精品视频免费| 国产精品一及| 别揉我奶头 嗯啊视频| 色噜噜av男人的天堂激情| 精品无人区乱码1区二区| ponron亚洲| 久久久久九九精品影院| 日韩,欧美,国产一区二区三区 | 三级国产精品片| 国产三级中文精品| 成人性生交大片免费视频hd| 三级国产精品欧美在线观看| 国产在视频线精品| 大话2 男鬼变身卡| 午夜免费激情av| 99视频精品全部免费 在线| 精品免费久久久久久久清纯| 亚洲在线自拍视频| 久久这里只有精品中国| 18禁动态无遮挡网站| 国产精品国产三级专区第一集| 大香蕉97超碰在线| 波多野结衣巨乳人妻| 97在线视频观看| 午夜视频国产福利| 男女啪啪激烈高潮av片| 乱系列少妇在线播放| 亚洲精品aⅴ在线观看| 国产片特级美女逼逼视频| 国产老妇女一区| 丰满少妇做爰视频| 国产精品1区2区在线观看.| 嫩草影院入口| 女人久久www免费人成看片 | 伊人久久精品亚洲午夜| 一级二级三级毛片免费看| 成人一区二区视频在线观看| 一边亲一边摸免费视频| 夜夜爽夜夜爽视频| 国产精品三级大全| 美女被艹到高潮喷水动态| 亚洲内射少妇av| 国产成人a区在线观看| 国产精品久久久久久精品电影| 国产视频内射| 亚洲欧美日韩无卡精品| 亚洲欧美清纯卡通| 夜夜看夜夜爽夜夜摸| 97在线视频观看| 午夜免费男女啪啪视频观看| 亚洲国产欧美人成| 国产成人一区二区在线| 99九九线精品视频在线观看视频| 免费看光身美女| 欧美性猛交╳xxx乱大交人| a级一级毛片免费在线观看| 国产一区亚洲一区在线观看| 亚洲在线观看片| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 人人妻人人澡欧美一区二区| 在线免费观看不下载黄p国产| 日日啪夜夜撸| 精品无人区乱码1区二区| 欧美三级亚洲精品| 精品一区二区三区视频在线| 亚洲av日韩在线播放| 简卡轻食公司| 国内少妇人妻偷人精品xxx网站| 午夜激情欧美在线| 国产乱人视频| 午夜精品在线福利| 两个人的视频大全免费| 欧美色视频一区免费| 日本免费在线观看一区| 能在线免费看毛片的网站| h日本视频在线播放| 国模一区二区三区四区视频| 成人亚洲欧美一区二区av| 国产亚洲av嫩草精品影院| 一个人看的www免费观看视频| av国产久精品久网站免费入址| 三级经典国产精品| 中文字幕熟女人妻在线| 成人美女网站在线观看视频| 99热网站在线观看| 中文字幕久久专区| 日本wwww免费看| 亚洲欧美一区二区三区国产| av在线观看视频网站免费| 午夜老司机福利剧场| 天堂av国产一区二区熟女人妻| 婷婷色综合大香蕉| 丰满少妇做爰视频| 国产又色又爽无遮挡免| 精品久久久久久久久亚洲| 亚洲av免费高清在线观看| 99久久九九国产精品国产免费| 日韩欧美三级三区| 成年免费大片在线观看| 99视频精品全部免费 在线| 校园人妻丝袜中文字幕| 亚洲在线自拍视频| 国产av在哪里看| av国产久精品久网站免费入址| 国产午夜精品一二区理论片| 午夜a级毛片| 亚洲国产最新在线播放| 一个人看视频在线观看www免费| 在线免费观看的www视频| 超碰97精品在线观看| 国产免费福利视频在线观看| 日本av手机在线免费观看| 成人综合一区亚洲| 亚洲精品成人久久久久久| 中文乱码字字幕精品一区二区三区 | 国产av不卡久久| 蜜桃久久精品国产亚洲av| 久久精品熟女亚洲av麻豆精品 | 又粗又硬又长又爽又黄的视频| 一级毛片电影观看 | 精品国内亚洲2022精品成人| 亚洲国产欧洲综合997久久,| 亚洲在线观看片| 久久久久久久久久黄片| 国产精品久久久久久久电影| 亚洲av.av天堂| 人妻系列 视频| 国产高清不卡午夜福利| 久久精品综合一区二区三区| 黄片无遮挡物在线观看| 天天躁日日操中文字幕| 看黄色毛片网站| 99久久精品热视频| 热99re8久久精品国产| 精品不卡国产一区二区三区| 欧美一区二区亚洲| 男女那种视频在线观看| 国内少妇人妻偷人精品xxx网站| 亚洲精品国产av成人精品| 亚洲成人av在线免费| 一区二区三区免费毛片| av天堂中文字幕网| 校园人妻丝袜中文字幕| 国产成人freesex在线| 日本免费a在线| 久久精品国产亚洲av涩爱| 国产成人福利小说| 日韩,欧美,国产一区二区三区 | 国产伦一二天堂av在线观看| 嫩草影院新地址| 91狼人影院| 日本av手机在线免费观看| 天天一区二区日本电影三级| 日韩av在线大香蕉| 日本免费在线观看一区| 国产精品嫩草影院av在线观看| 国产精品久久久久久久电影| 国产精品熟女久久久久浪| 91久久精品电影网| 久久草成人影院| 26uuu在线亚洲综合色| 久久精品国产亚洲av天美| 黄片wwwwww| 亚洲在线自拍视频| 九色成人免费人妻av| 国产又色又爽无遮挡免| 嫩草影院入口| 久久精品91蜜桃| 在线免费十八禁| 日本午夜av视频| 亚洲人成网站在线播| 久久99热这里只有精品18| 韩国av在线不卡| 日韩制服骚丝袜av| 亚洲人成网站高清观看| 久久久久久久午夜电影| 免费黄色在线免费观看| 黄片无遮挡物在线观看| 成年av动漫网址| 五月伊人婷婷丁香| 夜夜爽夜夜爽视频| 激情 狠狠 欧美| 日韩国内少妇激情av| 搡老妇女老女人老熟妇| 欧美日韩一区二区视频在线观看视频在线 | 秋霞在线观看毛片| 久久久成人免费电影| 国产成人精品一,二区| 观看美女的网站| 免费观看在线日韩| 一级av片app| 亚洲18禁久久av| 日韩国内少妇激情av| 欧美成人a在线观看| 国产精品久久久久久久电影| 韩国高清视频一区二区三区| 国产不卡一卡二| 中文在线观看免费www的网站| 中文字幕熟女人妻在线| 国内精品一区二区在线观看| 色播亚洲综合网| 网址你懂的国产日韩在线| 日日干狠狠操夜夜爽| 中文资源天堂在线| 亚洲精品乱久久久久久| 欧美成人午夜免费资源| 亚洲精品成人久久久久久| 国产精品美女特级片免费视频播放器| 国语自产精品视频在线第100页| 中文字幕熟女人妻在线| 丝袜喷水一区| 日韩中字成人| 少妇熟女欧美另类| 日本黄色视频三级网站网址| 午夜老司机福利剧场| 久久久久国产网址| 国产黄色小视频在线观看| 18+在线观看网站| АⅤ资源中文在线天堂| 九九爱精品视频在线观看| 黄片wwwwww| 久久精品久久久久久久性| 日韩一本色道免费dvd| 午夜老司机福利剧场| 国产亚洲精品av在线| 国产成年人精品一区二区| 国产乱人偷精品视频| 日韩一本色道免费dvd| 大香蕉97超碰在线| 免费av不卡在线播放| 日韩高清综合在线| 国产成人一区二区在线| 男女国产视频网站| 中文字幕av在线有码专区| 国产成人福利小说| 亚洲丝袜综合中文字幕| 中文精品一卡2卡3卡4更新| 久久久久久久久久久丰满| 国产精品av视频在线免费观看| 成人二区视频| 人人妻人人澡欧美一区二区| 51国产日韩欧美| 日本免费一区二区三区高清不卡| 国内精品宾馆在线| 亚洲电影在线观看av| 好男人视频免费观看在线| av天堂中文字幕网| 乱系列少妇在线播放| 国产午夜精品久久久久久一区二区三区| 精品99又大又爽又粗少妇毛片| 亚洲国产欧美在线一区| 久久久久久久久久久免费av| 国产人妻一区二区三区在| 精华霜和精华液先用哪个| 精品久久久久久久人妻蜜臀av| 成人午夜高清在线视频| 国产免费福利视频在线观看| 亚洲久久久久久中文字幕| 国产国拍精品亚洲av在线观看| 亚洲无线观看免费| 国产精品熟女久久久久浪| 久久99蜜桃精品久久| 级片在线观看| 一区二区三区四区激情视频| 国产一级毛片七仙女欲春2| 欧美3d第一页| 亚洲国产欧美人成| 白带黄色成豆腐渣| 国产真实乱freesex| 久久久久久伊人网av| 欧美变态另类bdsm刘玥| 69人妻影院| 天堂中文最新版在线下载 | 国产极品精品免费视频能看的| 成年免费大片在线观看| 九色成人免费人妻av| 美女脱内裤让男人舔精品视频| 韩国高清视频一区二区三区| 日韩大片免费观看网站 | 色吧在线观看| 亚洲人成网站在线观看播放| 午夜免费激情av| 欧美日本亚洲视频在线播放| 可以在线观看毛片的网站| 两个人视频免费观看高清|