• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    View suggestion for interactive segmentation of indoor scenes

    2017-06-19 19:20:13ShengYangJieXuKangChenandHongboFu
    Computational Visual Media 2017年2期

    Sheng YangJie Xu,Kang Chen,and Hongbo Fu

    View suggestion for interactive segmentation of indoor scenes

    Sheng Yang1Jie Xu2,Kang Chen1,and Hongbo Fu3

    Point cloud segmentation is a fundamental problem.Due to the complexity ofreal-world scenes and the limitations of3D scanners,interactive segmentation is currently the only way to cope with all kinds of point clouds.However,interactively segmenting complex and large-scale scenes is very time-consuming. In this paper,we present a novel interactive system for segmenting point cloud scenes.Our system automatically suggests a series of camera views,in which users can conveniently specify segmentation guidance.In this way,users may focus on specifying segmentation hints instead of manually searching for desirable views of unsegmented objects,thus significantly reducing user effort.To achieve this,we introduce a novelview preference model,which is based on a set of dedicated view attributes,with weights learned from a user study.We also introduce support relations for both graph-cut-based segmentation and finding similar objects.Our experiments show that our segmentation technique helps users quickly segment various types of scenes,outperforming alternative methods.

    point cloud segmentation;view suggestion; interactive segmentation

    1 Introduction

    With the prevalence of consumer-grade depth sensors(e.g.,Microsoft Kinect),scanning our living environments is becoming easier.However,the resulting 3D point clouds are often noisy, incomplete,and distorted,posing various challenges to traditional point cloud processing algorithms. Thus,in recent years,growing attention has been paid to low-quality point cloud processing problems. Amongst them,semantic segmentation,which aims to provide a decomposition of a 3D point cloud into semantically meaningful objects,is one of the most fundamental problems,and is important for many subsequent tasks such as object detection[1],object recognition[2],scene understanding[3],etc.

    Semantic segmentation of 3D point clouds has been extensively studied,resulting in various techniques,based for instance on region growing[4, 5],graph-cut[6–8],learning[9–11],etc.Most of those approaches attempted to achieve semantic segmentation with little or even no user intervention. However,due to the complexity of real-world scenes and the limitations of 3D scanners,manual intervention is often inevitable[12].

    Previous interactive segmentation work(e.g.,Refs. [13,14])typically focuses on improving segmentation results given the same amount of user input (e.g.,provided by a commonly used stroke-based interface).We observed that when interactively segmenting scenes at a moderate or large scale, finding appropriate views to provide segmentation hints is very time-consuming.For example,for a scene with multiple rooms containing objects of various types,shapes,and sizes,objects can easily occlude each other,requiring careful selection of viewpoints for interactive segmentation.In addition,due to the discrete nature of point clouds, the distances between viewpoint and objects need to be carefully chosen to ensure the desired point density and that contextual information is in view.

    Fig.1 Given an input scene represented as a 3D point cloud(left),our system automatically suggests a series of reasonable views(middle) for easily inputting segmentation hints for semantic segmentation of the entire scene(right).

    Based on these observations,we present a new interactive system for segmenting cluttered point clouds of large real-world scenes.Our system is able to automatically suggest a series of camera views,in which users can conveniently specify segmentation guidance,i.e.,2D strokes in our case.To reduce user effort,we aim to optimize the suggestions, i.e.,to provide views that both contain plenty of undetermined objects and can clearly display them. To achieve this,we introduce a novelview preference model,which is based on a set of dedicated view attributes,whose weights are learned from a user study.Given a new scene,our system uses the learned view preference model to find the next best views one by one.In this way,users may focus on specifying segmentation hints,instead of manually searching for desired views for segmentation of unsegmented objects.

    To further reduce user effort in interactive segmentation,we incorporate support relations in a graph-cut-based segmentation framework,to find similar objects for segmentation propagation. We have compared the performance of interactive point cloud segmentation with and without view suggestion,and interactive segmentation of RGB-D images.The experiments show that our segmentation technique with view suggestion helps users quickly segment various types of scenes,and outperforms alternative methods.

    2 Related work

    2.1 Point cloud segmentation

    Semantic segmentation of 3D point clouds or RGBD images has long been an active research topic in the communities of computer graphics,vision,and robotics(see an insightful survey by Nguyen and Le[12]).Below we discuss the most relevant works, and categorize them into supervised,unsupervised, and interactive techniques.

    With the growing availability of free 3D datasets (e.g.,theNYU Depth Dataset[11,15],theSUN3Ddataset[16],and ShapeNet[17]),supervisedlearning-based segmentation algorithms typically exploit high-level semantic information from labeled datasets,and use the learned knowledge to help detect,recognize,and thus semantically separate object regions from backgrounds.Training data for supervised methods mainly come from two sources: CAD models and RGB-D images.High-quality CAD models are ideal training data as they provide full 3D a priori knowledge about geometric shapes[10, 18]and even contextual relationships[19],but the number and diversity of high-quality digital models are far from enough to cover everything;creating these models is expensive.Since RGB-D images are much easier to acquire,numerous methods(e.g., Refs.[1,3,20,21])learn various features from labeled RGB-D images.However,despite ease of acquisition,the labeling of RGB-D images is still labor intensive.Our system can be used as a convenient and robust segmentation tool to help produce high-quality training data.

    Unsupervised semantic segmentation methods often rely on regular patterns(i.e.,symmetry and repetition)observed from input data itself,and work extremely effectively on outdoor building facades[22,23].However,interior scenes present much more complex structures.To simplify the problem,previous works(e.g.,Refs.[9,24])focus on large-scale working environments containing limited types of objects(e.g.,office desk,office chair, monitor),each repeated many times.However,realworld scenes often contain unique objects.Thus, manual intervention is often still required to refine segmentation results produced by such automatic methods.

    Since automatic segmentation methods are far from perfect,in practice interactive methods are more frequently used for segmenting both indoorscenes[13,25]and outdoor scenes[14].The most commonly used interactive scheme is to let users specify representative foreground and background regions(typically via a stroke-based interface),which are then used to construct a probabilistic inference model(using,e.g.,conditional random fields)and optimized using graph-cut[7,8],or simply used as seeds for region growing[4,5,26].However,such an interactive scheme is designed for segmenting an individual image frame,and thus requires carefully selected views of 3D scene contents for projection to screen space.While the view selection process is time-consuming,especially for large-scale and complex scenes,it has gained little attention.Thus, our work is largely complementary to existing interactive and automatic segmentation techniques.

    2.2 Camera control

    Finding feasible views to display 3Dgraphics on a 2D screen is a fundamental problem,which in computer graphics is generally referred to as the camera control problem[27,28].According to the camera mode and the scale of the 3D contents,existing solutions to this problem can be divided into two categories:fl yaroundandwalk-through.We adapt these existing ideas to a new context for interactive point cloud segmentation,and tackle unique challenges such as the handling of cluttered point clouds,and dynamic selection of a series of good views for easy labeling of objects.

    Afl y-aroundcamera allows complete contents rendered to the screen,and is often used when displaying single objects or small-scale scenes.The core problem is best view selection,which aims to automatically select the most representative view of a 3D model.Various low-level view attributes(e.g., projected area,viewpoint entropy,silhouette length, and depth distribution)have been proposed[29,30] as a basis for solving this problem.The state of the art is probably the work by Secord et al.[29],which learns how to combine low-level view attributes based on an extensive user study ofhuman preferences.Our work differs from the existing best view selection techniques in terms of both inputs(cluttered point clouds versus clean surface or volume models)and outputs(a series of viewpoints looking at different parts of a scene versus a set of independent good views looking at the same target).

    Walk-throughcamera mode is thus more relevant to our problem;it is often used to navigate within large-scale virtual scenes[31,32].This problem essentially comprises two sub-problems:viewpoint selection[30]and path planning[33,34].Unlike the criteria used for fly-around cameras,measuring the quality of viewpoints within a scene is mainly based onviewpoint entropy[30].Path planning is needed to avoid penetrating objects or walls and to present smooth scene roaming.In contrast, our problem demands essentially discrete views for labeling,although smooth transition is weakly considered.Again,existing walk-through methods often take clean scenes as input.Additionally,their extracted walk-through paths can be pre-generated, while in our case the selection of the next best view depends on the segmentation progress and needs to be determined on the fly.We have also found that top views are very usefulfor our application,but they are seldom used for the walk-through applications.

    3 System overview

    The main contribution of our work is automatic suggestion of good views for easy labeling of objects in indoor scenes represented as 3D point clouds.We implement view suggestion in an interactive segmentation system,which takes an unsegmented scan of an indoor scene as input. However,our interactive system can be helpful for interactively refining automatically generated segmentation results.

    As shown in Fig.2,our system contains three components:preprocessing,view suggestion,and interactive segmentation.In the preprocessing step (Section 4)the system automatically aligns an input scene with Manhattan directions,extracts storeys(levels),and clusters the points into patches which should be treated as indivisible units in segmentation.

    Afterwards our system automatically finds candidate viewpoints and sample views for presenting as suggestions(Section 5.1).To evaluate the quality ofa view,we introduce a view preference model(Section 5.2),which involves severalattributes such as point density,projected area,and viewpoint entropy.The weights of these attributes are learned by conducting a user study of human preferences for views of scenes.At runtime,each view is evaluated using the learned view preference model,withsegmentation status and smooth transitions taken into account.The best view is then suggested to the user for interactive segmentation.The user may reject the current suggestion,and our system then updates the strategy for suggesting another view.

    Fig.2 System overview.

    Given a view,the user provides segmentation hints on unsegmented,over-segmented,or undersegmented objects that are in view,using a classic stroke-based interface.The patches corresponding to these strokes are used as seeds to trigger a graph-cut-based segmentation optimization.We extract support relations between patches in the preprocessing step,and incorporate them into the segmentation optimization(Section 6.1).To further reduce the amount of user intervention,we find similar objects to already-segmented objects for segmentation propagation(Section 6.2).The above steps of automatic view suggestion and interactive segmentation ofscenes in the selected suggested view are repeated until all objects have been labeled. For multi-storey cases,our system provides view suggestions storey by storey.

    4 Preprocessing

    In this section we briefly introduce our preprocessing step,which generates essential information for later use.Our system takes as input a 3D point cloud of an indoor scene,which can be acquired using different types ofscanning devices such as LiDAR or Microsoft Kinect.For example,for RGB-D streams, they can be registered to form a point cloud using KinectFusion or its variants[35,36].Like many other point cloud processing pipelines[37],we downsample the data,estimate point normals, and transform the input scene into Manhattan coordinates[38]with thez-axis pointing upwards. Then we extract horizontal planes(by RANSAC) as floors and an optional ceiling.As illustrated in Fig.3,we decompose a multiple-storey building into individual storeys,with the ceiling of each storey removed to achieve top views of rooms.

    4.1 Constructing representative patches

    Over-segmentation has often been used in existing works for preprocessing point clouds[7,14,24].We are interested in over-segmenting an input scene into semantic patches,instead of patches with similar sizes but with no semantics[7].Such patches not only help reduce computational costs but also serve as integralsemantic units to analyze the importance of views and support relations.

    Fig.3 Decomposition of one-storey(left)and two-storey(right) buildings into individual storeys,with the removal of ceilings(if any).

    Our solution is an extension of the region growing approach by Ref.[24].Since the original approach does not consider any color information,it might not split objects with similar shape but different colors(e.g.,causing objects on the table in Fig.4(c) to disappear).In addition,their approach has been shown to be effective only on good-quality scenes acquired by LiDAR devices,and can lead to numerous tiny patches for point clouds of low quality(e.g.,those acquired by Kinect).To address these problems,we improve their approach by appropriately adding a new color condition and performing two rounds of growing,the first on the input point cloud and the second on the patches resulting from the first round.

    The first round of region growing is applied to the input points.Specifically,letG0=〈V0,E0〉denote a graph,withV0representing the input points andE0the edges generated byk-nearest neighborhoods (k=15 in our implementation).As points in different colors more likely belong to different objects or parts,as well as the basic normal and position conditions(with the default parameterst0andt1)for patch growing in Ref.[24],we add a color condition as follows:

    Fig.4 Patch construction results.(a)Downsampled point clouds from diff erent datasets.(b)By Ref.[7].(c)By Ref.[24].(d)Our result.

    whereiandjare two colors.[0,2π)and[0,1]denote normalized HSV color values fori.=(cossin)denotes the 2D coordinate values ofiin the Hue–Saturation color disk.Intuitively,we allow two colors to be merged when their Euclidean distance in the Hue–Saturation color disk and intensity value difference are both within a small thresholdt2(=0.05 in our implementation).

    This region growing process results in a set of merged patches.However,in our experiments we found that by adding the color condition,individual points might be isolated as individual patches, especially when the quality ofthe input point cloud is poor.To address this problem,we perform a second round of region growing on the merged patches from the first pass,but with relatively looser rules. LikeG0is constructed over points,G1=〈V1,E1〉is constructed over patches.For merging,we sort the patches in descending order of the number of contained points.For each patch,we find the normalof the best-fit plane and its centroid to check the respective normal and position conditions(the corresponding parameters are denotedt3,t4).The color condition is based on the color distribution and is more specifically defined using theχ-squared distance[39]between the HSV color histograms of two patchesPiandPjinV1:

    whereIidenotes the normalized HSV color histogram of points inPi.Specifically,H,S, and V channels are discretized into 16,16,and 8 bins respectively.We discretize V values into fewer bins to reduce the influence of different lighting conditions.The thresholdt5is set to 1.6 in our system.The above patch-based region growing procedure is repeated multiple times(5 times in our implementation).After each iteration,the thresholds are relaxed by 20%to merge small isolated patches into larger patches.In our system,we sett0=0.8,t1=0.05,t3=0.75,andt4=0.2 for allinput scenes.

    5 View suggestion

    In this section we first discuss view sampling, then perceptual assessment of views,and finally how to automatically suggest views for interactive segmentation.

    5.1 View sampling

    View sampling simplifies our problem,since it results in a discrete set of views for assessment.A view is basically determined by three vectors:the camera position,view direction,and up direction.

    Inspired by previous works on urban city reconstruction using both airborne and street-borne data[40],we consider two types of views:interior perspectivesandtop views.Top views provide a good summary of a scene or its parts,since objects often lie on flat surfaces(e.g.,tables,floors)and are thus easily separable from a top view.In contrast,interior perspectives enable a more detailed examination of a scene from a closer distance,and are useful for dealing with objects of small size or objects that are blocked in top views(e.g.,objects in bookshelves).Thus these two types of views are largely complementary to each other.Next we explain how we sample these two types of views.

    For interior perspectives,we fix the orientation of a view so that its up vector in image space is aligned with the upright orientation of a scene[29]. An interior perspective can then be characterized by (x,y,z,θ,φ),wherex,y,andzdefine the camera position,andθandφcontrol the view direction. A possible viewpoint should meet the following requirements:it must stay inside a room(above the floor,away from the walls,below the ceiling if any) and avoid hitting objects.Objects are approximately detected as columns using the column representation proposed by Fisher et al.[41].We then perform uniform sampling in the space of a storey excluding the space occupied by the columns to get possible camera positions(x,y,z),each ofwhich is associated with a view sphere parameterized by(θ,φ)(see Fig.2).Bothθandφare sampled every 15°,resulting in 266 views at each viewpoint.

    For every top view,its view direction is fixed and always points downward along the negativez-axis. Lean view directions are not used since objects in such resulting views would often be severely blocked by vertical walls.Hence,θis fixed atπfor all top views in our implementation.We thus use(x,y,z,φ) for mapping the camera position of a view and its up direction.We uncover the ceiling of a room or storey and uniformly sample the viewpoints and poses inside its bounding box.It can be easily seen that higher viewpoints produce wider-range views but fewer details.In the case of a multiple-storey building,we decompose it into multiple storeys (Section 4),and sample interior perspectives and top views for individual storeys.

    5.2 View preference model

    For simplicity here we assume no object has been segmented in a view;we discuss how to incorporate the segmentation status in Section 5.3.The simplified problem is similar to the problem of best view selection for a single object;the latter is often represented as a polygonal mesh.In contrast we need to deal with scenes of objects,represented as cluttered point clouds.We observe that good views have at least the following properties.First,objects in such views should be easy to recognize.Second, such views should contain as manyunsegmentedobjects as possible so that only a small set of views are needed to cover every object in a scene. These two properties are somewhat conflicting.For example,a bird’s eye view of a scene might include many objects in the view,but individual objects might not be easily recognizable.Thus there is no single definition to describe whether a given view is good or not.

    We explored various attributes to describe the quality of a view from different perspectives;some are based on previous work while others were designed by us.A key challenge is to evaluate the impact of each attribute when combining them together into a unified view preference model.This is achieved by learning from view preferences ofhuman viewers in a user study.Below we first describe the important attributes indicated by the user study and then give the details of the study itself.

    In a preprocessing step,we pre-compute asupplementary meshfrom the point cloud using a greedy projection triangulation algorithm[42]. This supplementary mesh greatly facilitates visibility checking.In addition,observing that people tend to focus more on objects located at the center ofa view, we add weights to pixels in thesupplementary meshview:each pixelpicorresponds to a projected point; its weight is defined aswi=1,wherediis the normalized Euclidean distance betweenpiand the center of the view,whileλwis a small factor to help emphasize the central areas(λw=0.2 in our system).

    Our attributes are:

    Apa:projected area.This is a rather basic attribute and has been proved effective in previous works like Ref.[29].LetP={Pn}denote the set ofvisiblepatches in a view.Visibility is checked using the pre-computed supplementary mesh.Apa, the sum of the projected areas of visible patches in view,is then defined as

    whereSnis the projected area of a visible patch, where each pixel is weighted by the focus-attention weightwi.

    Ave:viewpoint entropy.To estimate the richness ofa view,we follow V′azquez et al.[30],and calculate view entropy on the supplementary mesh.

    Apd:point density.Due to the discrete nature of our point clouds,the distance between viewpoint and objects must be carefully set to provide the desired point density and contextual information in the view.We perform such a view-related density measurement by comparing the difference of views between the point cloud and the supplementary mesh.Letbe a pixel of the point cloud view, andbe the weighted areas that the pixels in the point cloud view and the mesh view belong to,for the same patchi.Apdrepresenting the average density of patches in view is then defined as

    where?(x)=1/(1+e?λ?·(x?t6))is a sigmoid function for evaluating the density of one patch. Intuitively,ifS′n/Snexceeds a thresholdt6,the density is acceptable and allows users to recognize objects.In our system,we setλ?=16 andt6=0.3.

    Besides the attributes listed above,we also explored additional attributes considering aesthetic perception,including:depth information(Add,Adc),object layout(Aop,Aod,Aov),andradial patches(Arp).Please refer to the Electronic Supplementary Material(ESM)for detailed formulations of these attributes.However,according to the user study, they are relatively dispensable.

    User study.In order to analyze the importance of these attributes to our view preference model, we conducted a user study.We chose three scenes captured by different types ofdevices(Kinect V2 and LiDAR)to alleviate the influence of diff erent input qualities.For each scene,we randomly sampled 60 views(using the sampling approach described in Section 5.1),including both top views(5%)and interior perspectives(95%).16 participants with basic knowledge ofcomputer graphics were recruited to manually evaluate the quality of each view.To ensure a proper comparative evaluation,each time, each participant was given a group ofviews from the same scene(6 views per group,randomly selected from 60 views)and was asked to rate each view on a scale from 1(worst)to 5(best).They were also told that views containing easily recognizable and multiple objects without severe occlusion should be given high scores.Representative views are shown in Fig.5.Please refer to the ESM for other scenes and results.

    Learning weights of attributes.After collecting the scores for each view,we studied the importance of attributes.We normalized the observed attributes and adopted lasso regression[43] to determine the weightsβiof the attributes:it isable to perform variable selection in order to enhance the accuracy of the predictions by the statistical model it produces.The learned model by lasso regression is simply formulated as

    Fig.5 Representative views from the user study.(a)High-score top view.(b,c)High-score interior views.(d)Low-score interior view(the tree seriously blocks the objects behind).(e,f)Low-score interior views.

    Lasso regression shows that among all the attributes included in the regression testing,view entropyandpoint densityplayed the most significant roles,while other attributes were dispensable.Thus, we discarded these other attributes and performed lasso regression again.The final weights provided forview entropyandpoint densitywere 2.11 and 3.05,respectively.Detailed information about the regression process can be found in the ESM.

    5.3 View suggestion

    Given an input scene,we first sample views(see Section 5.1)in the preprocessing step.At runtime, we use the learned view preference model(Eq.(5)) to automatically suggest views.Specifically,we generate a set of candidate views,denoted byV,by picking views withApa>0.6.The view inVwith the highest scoreis suggested as the first view.

    Given a suggested view,the user may perform interactive segmentation(see Section 6)and then request another view.To avoid repetitive suggestions,each time a view is suggested,we lower its score by a scale factor 0.7.We also allow users to reject a current view.In this case our system will suggest new views.We suggest new views according to the following principles:

    ?View suggestions should respect the current segmentation status.Views with many alreadysegmented objects should have lower priority.

    ?A smooth transition between the current and the new views is preferred.

    ?View suggestion should take into account user rejection of suggested views.

    To satisfy the first guideline,we update the score of each candidate view according to the current segmentation status.More specifically,we replaceAvewith the residual entropywhich is calculated by removing the terms corresponding to already-labeled patches inAve,since such labeled patches provide little information.In contrast,Apdremains unchanged because intuitively this term is used to clearly display point clouds to the screen, and thus is somewhat independent of labeling.

    To encourage a smooth transition fromVntoVn+1, we pickVn+1according to the following rule:

    whereis the online score of viewVx,andD(Ki,Kn)is the horizontal Euclidean distance between the current camera positionKnand the candidate positionKi.We use an exponentialdecay ofpreference to favor nearby rather than distant view candidates.

    If a user rejects a suggested view,it generally means that unlabeled patches in this view are less important,and thus,their priority should be lowered when assessing subsequent suggestions.To do so,we lower the residual entropyAveof unlabeled patches in rejected views by a scale factor 0.8.

    For multi-storey cases,our system begins with the highest storey and suggests views related to the current storey until the labeling progress of the current storey surpasses 90%.Once surpassed,our system will suggest the best view for the next lower storey and continue.

    6 Interactive segmentation

    We adopt a stroke-based interface similar to that in Ref.[14]for interactive segmentation.Given a view,users draw different strokes on unsegmented regions to indicate different objects(see Fig.6(left)). Graph-cut-based optimization[44]is then used to achieve the desired segmentation results.Our work contributes to this paradigm by introducing support relations into graph-cut,and finding similar objects to further reduce the amount of user intervention.

    6.1 Support relations for graph-cut

    Fig.6 Graph-cut without(middle)and with(right)support relations.Left:input strokes on unsegmented objects in a given view.

    Previous graph-cut-based point cloud segmentation techniques(e.g.,Refs.[7,14])mainly focus on theuse of low-level geometric features(e.g.,position, normal,color)to propagate the segmentation information from the user-specified seeds to the rest of the scene.Since man-made objects often exhibit box-like shapes,due to the sharp change in normal between different faces of such objects, multiple strokes(one for each face)may be needed to segment a single object(e.g.,a desk or monitor in Fig.6(middle)).

    To reduce the amount of user intervention, we introducesupport relationsas an additional measurement of distance between patches in the graph-cut formulation.This helps the reduction of both the number of user-specified strokes and the number of needed views.For example,with our support relations considered(see Fig.6(right)),the user only needs to input strokes on the top faces of a desk and monitor in a top view to segment the objects.Without,the user would need to switch to different views to input further strokes.

    Support relations are common in indoor scenes because of the influence of gravity.There are two levels of support relations:support between objects and support between primitive shapes which constitute an object.The former requires the availability of object-level segmentation(the output ofour system)[11].Our focus is on the second type of support relations to guide object-levelsegmentation.

    We take the patches constructed in Section 4.1 as input.First we classify all patches into two types:quasi-verticalQvand quasi-horizontalQhaccording to the angles between the planes fitted to them and the ground plane(relative to 45°). Then we concentrate on two common relationships: a quasi-verticalpatch supporting an quasi-horizontal patchQvhabove it,and a quasi-vertical patch being supported by a quasi-horizontal patchQhvbelow it, as shown in Fig.7.

    For each pair of an adjacent quasi-vertical patchPiand a quasi-horizontal patchPj(abovePi), the likelihood that patchesPiandPjhave aQvhrelationship is defined as follows:

    Fig.7 Automatically detected support relations.Left:the quasivertical patches co-support the top quasi-horizontal patches.

    whereU(Pi,Pj)is the number of quasi-vertical patches within the neighborhood ofPi(includingPi)which havePjas an upper patch andλQ=0.2 is a fixed parameter.Intuitively,more quasi-vertical patches supporting the same quasi-horizontal patch lead to higher likelihood of having theQvhrelationship.

    Similarly,for each pair of an adjacent quasivertical patchPiand a quasi-horizontal patchPj(belowPi),the likelihood that patchesPiandPjhave aQhvrelationship is defined as follows:

    whereWiis the area of the convex hull of points in patchPi.We consider aQhvrelationship for a quasi-vertical patchPionly when there is noQvhrelationship involvingPi.Consider Fig.7(middle): the laptop lid hasQhvrelationship with the table, the laptop base,and the keyboard,while the laptop base leads to the highest likelihood since they have about the same size.

    Then we defineTS(Pi,Pj),the likelihood for patchesPiandPjto belong to the same object using inference from the support map,as follows:

    whereQxcan be eitherQvhorQhv,decided by the spatial relationship between two concerning patches.The top two conditions are essentiallyQvhorQhv,respectively,and the third indicates cosupport relations between patches inQv,which exist ifand only iftwo verticalpatches have the same type of support relations(Qx)with the same horizontal patch(see the example in Fig.7(left)).Since support relations between patches are irrelevant to the segmentation status,in our system,they are computed during the preprocessing step.

    Graph-cut.Given a set of user-specified strokes, only pixels with the same patch indices on both the point cloud view and the supplementary mesh view are regarded as valid,to avoid penetrationlabeling.We then formulate support relations as well as distances,colors,and normals in the graph-cut formulation.Please refer to the ESM for details.

    6.2 Finding similar objects

    Duplicate objects often exist in a scene.It is redundant for a user to label them one by one.Our system employs a simple algorithm to automatically retrieve candidate objects similar to a given object in the scene.Ifthe found objects are in fact not similar, users can delete the automatically generated labels.

    Several methods for finding similar objects in point cloud scenes have already been proposed.For example,Kim et al.[9]introduced a method by first learning models of frequently occurring objects and then performing real-time matching.Mattausch et al.[24]took patch similarity and spatialconsistency into account and automatically found all similar objects in a scene by clustering,which,however, takes dozens of seconds for large-scale indoor scenes.

    We aim for a simple and efficient tool that can perform in real time,without pre-trained models or clustering.Our approach is based on the key assumption that similar objects should contain similar support structures,i.e.,both the support relations and the accompanying patches should be similar.This assumption allows us to identify similar objects by straightforwardly comparing allunlabeled candidate support structures with a given object, based on patch similarity and support relations.

    More specifically,given an already segmented object,denoted byC0,we first identify allcandidate patches which have not been segmented and have a high similarity(t7=0.3 in our implementation)to at least one patch inC0.For efficiency,we slightly change the similarity metric in Ref.[24]to define our similarityAi,jfor a given pair of patchesPiandPj(see the ESM).

    Then we group these candidate patches by connectedness to form patch constellations.For each pair of patches〈Pi,Pj〉in each constellationCn,we search for a corresponding pair〈Pk,Pl〉∈C0with the same type of support relation.Based on our key assumption,we compute the similarity likelihood Sim(Cn,C0)betweenCnandC0as follows:

    If the similarity likelihood Sim(Cn,C0)exceeds a thresholdt8(=0.25 in our implementation),we take it as a candidate similar object.We then use all patches in the found isomorphic support structures as seeds for a new object to perform another round of graph-cut.

    As shown in Fig.8 our algorithm manages to find several similar objects in two cases.In the left example,each chair is composed of different numbers of patches,but the algorithm is still able to find such similar isomorphic parts and perform graph-cut.The right example shows our ability to find diff erent types of objects:tables (green),monitors(cyan),telephones(yellow),and keyboards(purple).Note that objects adjacent to some keyboards are mistaken by graph-cut, demanding additionalstrokes for refinement.

    7 Results and evaluation

    We have tested our system on different qualities of datasets from different sources.We also performed a user study to evaluate the performance of view suggestion.Please see the accompanying video in the ESM for interactive demos.Some representative suggested views and segmentation results are shown in Fig.9.

    Datasets.In order to test our performance of view suggestion and segmentation,we chose middle-or large-scale scenes with rich sets ofobjects. Specifically,we used 2 out of 3 scenes(office2, office3)in the room detection datasets of UZH[24], 1 of 5 scenes(office1)in the Floored Panorama dataset[37],and 2 of 2 synthetic scenes(officeroom,living-room)with ground-truth trajectory in the ICL-NUIM dataset[45].The ICL-NIUM datawas generated by sampling point on surfaces of existing 3D mesh-based models,while others were acquired with LiDAR(with color information).Figure 9 illustrates all scenes tested.We additionally collected a scene captured by Microsoft Kinect2 to test for poor-quality scans.

    Fig.8 Results of finding similar objects(shown in the same color).

    Fig.9 Examples of results of our system.Left:view score map.Right:high score examples when no objects are labeled.

    Fig.10 Results of user study.Top,left:scenes with pre-defined target segmentations.(a)Acceptance rate of suggestions and hypothesis test results.Thep-values are from paired samplet-tests between our method and no suggestion(n)or RGB-D(d),giving full time(T),the time spent on manual navigation(N),the counts of minimap navigation(M)and used views(V).Right:average and standard deviation of the means of observed values.(b)Time spent on the whole process in second.(c)Time spent on manual navigation in second.(d)The count of used views.

    7.1 Validity of view suggestion

    To evaluate the effectiveness of view suggestion, we compared the performance of interactive segmentation with and without view suggestion.We also compared to stroke-based segmentation directly on RGB-D frames ifavailable.These three methods, denoted as M-S,M-N,and M-D,were compared in a user study.Please refer to the accompanying video in the ESM to see how each method works.

    Interfaces.For each of the methods compared, we provided a minimap for easy navigation.For M-D,while interactive segmentation was done on RGB-D images,we provided an additional point cloud viewer for examining the segmentation status.The only difference between M-S and MN was the availability of view suggestion.For MS,we encouraged the participants to use the view suggestion feature as much as possible,though minimap navigation and common 3D browsing were still allowed.All methods were tested in fullscreen mode(1080 p)to eliminate the influence of window size.For fair comparison,we used the same preprocessing,graph-cut,and similar-object-finding algorithms for all methods.

    Participants and apparatus.We recruited 12 university students,who had basic knowledge of graphics but different levels of experience in 3D browsing.In a training session,which lasted nearly 15 minutes on average,each participant was briefed and trained how to use the three methods.The entire user study was conducted on the same PC with i7 2.10 GHz CPU,a GTX 660Ti GPU,and 16 GB RAM.

    Tasks.We prepared 3 scenes with target segmentations.Scene A in Fig.10(a)came from the ICL-NUIM dataset with 42 objects and 35 RGB-D key-frames.Scene B was our own recorded Kinect2 data,including 22 objects and 33 keyframes.Scene C was a large-scale multiple-room scene with 21 objects but no associated RGBD images.These scenes are given in rows 1–3, respectively,in Fig.9.The participants were asked to interactively segment these scenes using the three methods,whose presentation was counterbalanced to alleviate bias due to familiarity with the scenes.For Scene C,we tested M-S and M-N only,due to its lack of RGB-D frames.We pre-defined a groundtruth segmentation for each scene with manually labeled objects and required the participants to reproduce the ground-truth segmentations(shown on another monitor simultaneously)as closely as possible.Since it was difficult to reproduce the segmentations exactly,each scene was considered as completed ifthe similarity(ratio ofpatches)between the current and target segmentations reached 90%.

    Measures.Our system recorded the following information for comparison:time spent on manual navigation,processing,and the whole progress (including planning,drawing,processing,and navigation);number of uses of view suggestion and minimap navigation operations;number of views used;number of executions ofgraph-cut and similarobject-finding algorithms;number and total length of input strokes.Additionally,we also recorded the acceptance ratio of view suggestion.We conducted a questionnaire survey at the end of the study.

    Results.The acceptance rates of our suggestedviews for Scenes A,B,and C were 83.9%,85.8%, and 93.1%,respectively,indicating the good quality of suggested views.Figures 10(b)–10(d)show the statistics of some observed values.The table in Fig.10(a)gives evaluation and hypothesis testing results,wherepis thep-value calculated through student’st-test[46],which is helpful for evaluating statistically significant differences among small-scale paired-samples.

    Using one-way analysis of variance(ANOVA),we found a significant difference in the average total time between the tested methods(p=0.0288 for all methods on Scenes A and B,p=0.0058 for M-S and M-N on all the scenes).In all tested scenes,our method significantly outperformed M-D (p<0.005 fromt-test)in both the number of views used and the total time.In Scene C,our method was significantly faster than M-N(p<0.001), although they achieved comparable performance on minimap navigation(p=0.0527).We observed that minimap navigation is crucial for large-scale scenes.In Scenes A and B,compared to M-N,our method significantly reduced the time for camera manipulation(p<0.005)and the usage of minimap navigation(p<0.03),but achieved comparable performance in total time(p>0.05).This might be because the expense ofmanualnavigation in such single-room scenes is relatively low.Our method also used significantly fewer views than M-N(p<0.02) when segmenting all the tested scenes.

    Our participants reported that with M-D,the RGB-D image views were more useful for object recognition due to their dense views,but(re-) checking the labeled objects,either through backand-forth traversal or with the help of the auxiliary point cloud view,was not convenient.Worse,the number of RGB-D key-frames can significantly increase as scenes get bigger.Some participants also reported that they subconsciously referred to the good RGB-D views or results of view suggestion to improve labeling efficiency when using M-N.In fact, the performance ofinteractive segmentation without view suggestion can be easily affected by familiarity with the scenes and the proficiency of 3D browsing.

    7.2 Robustness and efficiency

    Parameters.In the preprocessing step for extracting storeys and sampling viewpoints,we used different sets of parameters for diff erent datasets due to their different data quality and scene scale. The other parameters for patch construction,view suggestions,and interactive segmentation,remained unchanged for allthe tested scenes.

    Timing.On the same machine used in the user study,the preprocessing stage can cost dozens of minutes for the tested scenes.For a 140 m2onefloor room with 16 million points,it took about 1 minute per viewpoint to traverse through scenes to get the descriptors and 30 minutes in total;this could potentially be made faster by switching to quicker rendering tools.Since the score of each view can be calculated separately,we can also easily accelerate the method by parallelization.In the online stage, it only cost 50 ms to suggest an appropriate view, the graph-cut process only took 100 ms,and finding similar objects took 200 ms with respect to a labeled object.In summary,the online processing costs meet the requirement of real-time response.

    Limitations.In some scenes with complicated objects such as plants or trees,our system may“prefer”to suggest views involving such objects (Fig.11(left)),since they are taken as fragmentary patches(each including several leaves or stems)in patch construction,leading to high residual entropy on these views.However,once such objects are labeled,their influence willdisappear.Since patches are indivisible in our implementation,if two or more objects are gathered into one patch(e.g.,due to similar color and depth),it is impossible to segment the individualobjects.Also the boundaries ofobjects may not be accurate if the points on boundaries are ambiguous.This might be addressed by introducing an interactive repair tool for boundary refinement. In addition,our proposed support relations and the algorithm for finding similar objects may fail if objects are combined as multiple discrete components(see Fig.11(right)).Fortunately,users can modify the labeling by providing additionalinputs for segmentation.

    Fig.11 Less successful cases.Left:an example view with a tree with a high score.Right:failure to find similar objects in isolated patches(the chair in red).

    8 Conclusions and future work

    We have presented a novel system for interactive segmentation of large-scale indoor scenes, represented as cluttered point clouds.The key contributions of our work are the problem and solution of automatic view suggestion for interactive segmentation.Other contributions include support relations for graph-cut-based segmentation and finding similar objects.Our extensive evaluations show the advantages ofour approach over alternative methods.In future,besides addressing the discussed limitations,we are also interested in applying a similar idea ofview suggestion for other applications, e.g.,interactive editing of large-scale scenes.

    Acknowledgements

    This work was supported by the Joint NSFC–ISF Research Program(Project No.61561146393), the National Natural Science Foundation of China(Project No.61521002),the Research Grant of Beijing Higher Institution Engineering Research Center,and the Tsinghua–Tencent Joint Laboratory for Internet Innovation Technology. The work was partially supported by grants from the Research Grants Council of the Hong Kong Special Administrative Region,China(Project Nos. CityU113513 and CityU11300615).

    Electronic Supplementary Material Supplementary material is available in the online version of this article at http://dx.doi.org/10.1007/s41095-017-0078-4.

    [1]Lai,K.;Bo,L.;Ren,X.;Fox,D.Detection-based object labeling in 3D scenes.In:Proceedings of the IEEE International Conference on Robotics and Automation,1330–1337,2012.

    [2]Johnson,A.E.;Hebert,M.Using spin images for efficient object recognition in cluttered 3D scenes.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.21,No.5,433–449,1999.

    [3]Zheng,B.;Zhao,Y.;Yu,J.C.;Ikeuchi,K.;Zhu, S.C.Beyond point clouds:Scene understanding by reasoning geometry and physics.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,3127–3134,2013.

    [4]Holz,D.;Behnke,S.Fast range image segmentation and smoothing using approximate surface reconstruction and region growing.In:Intelligent Autonomous Systems 12.Lee,S.;Cho,H.;Yoon, K.-J.;Lee,J.Eds.Springer Berlin Heidelberg,61–73, 2013.

    [5]Rabbani,T.;van den Heuvel,F.A.;Vosselmann, G.Segmentation of point clouds using smoothness constraint.International Archives of Photogrammetry, Remote Sensing and Spatial Information SciencesVol. 36,No.5,248–253,2006.

    [6]Boykov,Y.;Funka-Lea,G.Graph cuts and efficient N-D image segmentation.International Journal of Computer VisionVol.70,No.2,109–131,2006.

    [7]Golovinskiy,A.;Funkhouser,T.Min-cut based segmentation of point clouds.In:Proceedings of the IEEE 12th International Conference on Computer Vision Workshops,39–46,2009.

    [8]Sedlacek,D.;Zara,J.Graph cut based pointcloud segmentation for polygonal reconstruction.In:Advances in Visual Computing.Bebis,G.;Boyle,R.; Parvin,B.;Koracin,D.et al.Eds.Springer Berlin Heidelberg,218–227,2009.

    [9]Kim,Y.M.;Mitra,N.J.;Yan,D.-M.;Guibas,L. Acquiring 3D indoor environments with variability and repetition.ACM Transactions on GraphicsVol.31, No.6,Article No.138,2012.

    [10]Nan,L.;Xie,K.;Sharf,A.Asearch-classifyapproach for cluttered indoor scene understanding.ACM Transactions on GraphicsVol.31,No.6,Article No.137,2012.

    [11]Silberman,N.;Hoiem,D.;Kohli,P.;Fergus,R. Indoor segmentation and support inference from RGBD images.In:Computer Vision–ECCV 2012. Fitzgibbon,A.;Lazebnik,S.;Perona,P.;Sato,Y.; Schmid,C.Eds.Springer Berlin Heidelberg,746–760, 2012.

    [12]Nguyen,A.;Le,B.3D point cloud segmentation:A survey.In:Proceedings of the 6th IEEE Conference on Robotics,Automation and Mechatronics,225–230, 2013.

    [13]Shao,T.;Xu,W.;Zhou,K.;Wang,J.;Li,D.; Guo,B.An interactive approach to semantic modeling of indoor scenes with an RGBD camera.ACM Transactions on GraphicsVol.31,No.6,Article No. 136,2012.

    [14]Yuan,X.;Xu,H.;Nguyen,M.X.;Shesh,A.; Chen,B.Sketch-based segmentation of scanned outdoor environment models.In:Proceedings of the Eurographics Workshop on Sketch-Based Interfaces and Modeling,19–26,2005.

    [15]Silberman,N.;Fergus,R.Indoor scene segmentation using a structured light sensor.In:Proceedings of the IEEE International Conference on Computer Vision Workshops,601–608,2011.

    [16]Xiao,J.;Owens,A.;Torralba,A.SUN3D:A database of big spaces reconstructed using SfM and object labels.In:Proceedings of the IEEE International Conference on Computer Vision,1625–1632,2013.

    [17]Chang,A.X.;Funkhouser,T.;Guibas,L.;Hanrahan, P.;Huang,Q.;Li,Z.;Savarese,S.;Savva,M.;Song, S.;Su,H.;Xiao,J.;Yi,L.;Yu,F.ShapeNet:An information-rich 3D model repository.arXiv preprintarXiv:1512.03012,2015.

    [18]Hinterstoisser,S.;Lepetit,V.;Ilic,S.;Holzer,S.; Bradski,G.R.;Konolige,K.;Navab,N.Model based training,detection and pose estimation of texture-less 3D objects in heavily cluttered scenes.In:Computer Vision–ACCV 2012.Lee,K.M.;Matsushita,Y.;Rehg, J.M.;Hu,Z.Eds.Springer Berlin Heidelberg,548–562,2012.

    [19]Chen,K.;Lai,Y.-K.;Wu,Y.-X.;Martin,R.;Hu,S.-M.Automatic semantic modeling ofindoor scenes from low-quality RGB-D data using contextualinformation.ACM Transactions on GraphicsVol.33,No.6,Article No.208,2014.

    [20]Silberman,N.;Sontag,D.;Fergus,R.Instance segmentation ofindoor scenes using a coverage loss.In:Computer Vision–ECCV 2014.Fleet,D.;Pajdla,T.; Schiele,B.;Tuytelaars,T.Eds.Springer International Publishing,616–631,2014.

    [21]Chen,K.;Lai,Y.K.;Hu,S.-M.3D indoor scene modeling from RGB-D data:a survey.Computational Visual MediaVol.1,No.4,267–278,2015.

    [22]Shen,C.-H.;Huang,S.-S.;Fu,H.;Hu,S.-M.Adaptive partitioning of urban facades.ACM Transactions on GraphicsVol.30,No.6,Article No.184,2011.

    [23]Zhang,H.;Xu,K.;Jiang,W.;Lin,J.;Cohen-Or, D.;Chen,B.Layered analysis of irregular facades via symmetry maximization.ACM Transactions on GraphicsVol.32,No.4,Article No.121,2013.

    [24]Mattausch,O.;Panozzo,D.;Mura,C.;Sorkine-Hornung,O.;Pajarola,R.Object detection and classification from large-scale cluttered indoor scans.Computer Graphics ForumVol.33,No.2,11–21,2014.

    [25]Valentin,J.;Vineet,V.;Cheng,M.-M.;Kim,D.; Shotton,J.;Kohli,P.;Nie?ner,M.;Criminisi, A.;Izadi,S.;Torr,P.SemanticPaint:Interactive 3D labeling and learning at your fingertips.ACM Transactions on GraphicsVol.34,No.5,Article No. 154,2015.

    [26]Wong,Y.-S.;Chu,H.-K.;Mitra,N.J.SmartAnnotator an interactive tool for annotating indoor RGBD images.Computer Graphics ForumVol.34,No.2,447–457,2015.

    [27]Christie,M.;Olivier,P.Camera control in computer graphics:Models,techniques and applications.In: Proceedings of the ACM SIGGRAPH ASIA 2009 Courses,Article No.3,2009.

    [28]Scott,W.R.;Roth,G.;Rivest,J.-F.View planning for automated three-dimensionalobject reconstruction and inspection.ACM Computing SurveysVol.35,No. 1,64–96,2003.

    [29]Secord,A.;Lu,J.;Finkelstein,A.;Singh,M.;Nealen, A.Perceptual models of viewpoint preference.ACM Transactions on GraphicsVol.30,No.5,Article No. 109,2011.

    [30]V′azquez,P.-P.;Feixas,M.;Sbert,M.;Heidrich, W.Viewpoint selection using viewpoint entropy.In: Proceedings of the Vision Modeling and Visualization Conference,273–280,2001.

    [31]And′ujar,C.;V′azquez,P.;Fair′en,M.Way-Finder: Guided tours through complex walkthrough models.Computer Graphics ForumVol.23,No.3,499–508, 2004.

    [32]Li,T.-Y.;Lien,J.-M.;Chiu,S.-Y.;Yu,T.-H. Automatically generating virtual guided tours.In: Proceedings of the Computer Animation,99–106, 1999.

    [33]Christie,M.;Langu′enou,E.A constraint-based approach to camera path planning.In:Smart Graphics.Butz,A.;Kr¨uger,A.;Olivier,P.Eds. Springer Berlin Heidelberg,172–181,2003.

    [34]Salomon,B.;Garber,M.;Lin,M.C.;Manocha,D. Interactive navigation in complex environments using path planning.In:Proceedings of the Symposium on Interactive 3D Graphics,41–50,2003.

    [35]Choi,S.;Zhou,Q.-Y.;Koltun,V.Robust reconstruction of indoor scenes.In:Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,5556–5565,2015.

    [36]Newcombe,R.A.;Izadi,S.;Hilliges,O.;Molyneaux, D.;Kim,D.;Davison,A.J.;Kohli,P.;Shotton,J.; Hodges,S.;Fitzgibbon,A.KinectFusion:Real-time dense surface mapping and tracking.In:Proceedings of the 10th IEEE International Symposium on Mixed and Augmented Reality,127–136,2011.

    [37]Ikehata,S.;Yang,H.;Furukawa,Y.Structured indoor modeling.In:Proceedings of the IEEE International Conference on Computer Vision,1323–1331,2015.

    [38]Furukawa,Y.;Curless,B.;Seitz,S.M.;Szeliski, R.Manhattan-world stereo.In:Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition,1422–1429,2009.

    [39]Asha,V.;Bhajantri,N.U.;Nagabhushan,P. GLCM-based chi-square histogram distance for automatic detection of defects on patterned textures.International Journal of Computational Vision and RoboticsVol.2,No.4,302–313,2011.

    [40]Fr¨uh,C.;Zakhor,A.Constructing 3D city models by merging aerial and ground views.IEEE Computer Graphics and ApplicationsVol.23,No.6,52–61,2003.

    [41]Fisher,M.;Savva,M.;Li,Y.;Hanrahan,P.;Nie?ner, M.Activity-centric scene synthesis for functional 3D scene modeling.ACM Transactions on GraphicsVol. 34,No.6,Article No.179,2015.

    [42]Marton,Z.C.;Rusu,R.B.;Beet,M.On fast surface reconstruction methods for large and noisy point clouds.In:Proceedings of the IEEE International Conference on Robotics and Automation,3218–3223, 2009.

    [43]Tibshirani,R.Regression shrinkage and selection via the lasso.Journal of the Royal Statistical SocietyVol. 58,No.1,267–288,1996.

    [44]Boykov,Y.;Veksler,O.;Zabih,R.Fast approximate energy minimization via graph cuts.IEEE Transactions on Pattern Analysis and Machine IntelligenceVol.23,No.11,1222–1239,2001.

    [45]Handa,A.;Whelan,T.;McDonald,J.;Davison, A.J.A benchmark for RGB-D visual odometry, 3D reconstruction and SLAM.In:Proceedings of the IEEE International Conference on Robotics and Automation,1524–1531,2014.

    [46]Gosset,W.S.The probable error of a mean.BiometrikaVol.6,No.1,1–25,1908.

    Jie Xu is a Ph.D.student at the Computer Science and Artificial Intelligence Laboratory in Massachusetts Institute of Technology. His research interests include computer graphics and geometric processing.

    Kang Chen received his B.S.degree in computer science from Nanjing University in 2012.He is currently a Ph.D.candidate in the Institute for Interdisciplinary Information Sciences, Tsinghua University.His research interests include computer graphics, geometric modeling and processing.

    Hongbo Fu is an associate professor in the School of Creative Media,City University of Hong Kong.He received his Ph.D.degree in computer science from the Hong Kong University of Science and Technology in 2007 and B.S.degree in information sciences from Peking University in 2002.His primary research interests fall in the fields of computer graphics and human computer interaction.He has served as an associate editor ofThe Visual Computer,Computers&Graphics,andComputer Graphics Forum.

    Open Access The articles published in this journal are distributed under the terms of the Creative Commons Attribution 4.0 International License(http:// creativecommons.org/licenses/by/4.0/),which permits unrestricted use,distribution,and reproduction in any medium,provided you give appropriate credit to the original author(s)and the source,provide a link to the Creative Commons license,and indicate if changes were made.

    Other papers from this open access journalare available free of charge from http://www.springer.com/journal/41095. To submit a manuscript,please go to https://www. editorialmanager.com/cvmj.

    g

    his B.S.degree in computer science from Wuhan University in 2014.He is currently a Ph.D.candidate in computer science in Tsinghua University.His research interests include computer graphics and point cloud processing.

    1 Tsinghua University,Beijing,China.E-mail:S. Yang,shengyang93fs@gmail.comK.Chen, chenkangnobel@hotmail.com.

    2 Massachusetts Institute of Technology,Cambridge,USA. E-mail:eternal answer@126.com.

    3 City University of Hong Kong,Hong Kong,China.E-mail:fuplus@gmail.com.

    Manuscript received:2016-12-02;accepted:2017-01-12

    亚洲精品中文字幕在线视频| 久热这里只有精品99| 久久韩国三级中文字幕| 日日摸夜夜添夜夜添av毛片| 久久久精品94久久精品| 成人午夜精彩视频在线观看| 国国产精品蜜臀av免费| 美女中出高潮动态图| 又粗又硬又长又爽又黄的视频| 精品久久久噜噜| 大片电影免费在线观看免费| 大又大粗又爽又黄少妇毛片口| 一本—道久久a久久精品蜜桃钙片| 精品亚洲成国产av| 国产一区亚洲一区在线观看| 欧美另类一区| 成人漫画全彩无遮挡| 少妇的逼水好多| 欧美+日韩+精品| 国产精品.久久久| 国产成人精品无人区| 精品熟女少妇av免费看| 母亲3免费完整高清在线观看 | 22中文网久久字幕| a级毛片在线看网站| 人人妻人人澡人人看| 国产亚洲午夜精品一区二区久久| 少妇丰满av| av福利片在线| 久久99精品国语久久久| 午夜av观看不卡| 日日撸夜夜添| 婷婷色综合大香蕉| 亚洲av.av天堂| 中国三级夫妇交换| 2018国产大陆天天弄谢| 成人漫画全彩无遮挡| 嫩草影院入口| videossex国产| 五月天丁香电影| 熟女人妻精品中文字幕| 亚洲五月色婷婷综合| 亚洲欧美日韩另类电影网站| 少妇精品久久久久久久| 久久国产精品大桥未久av| 亚洲精品久久久久久婷婷小说| av又黄又爽大尺度在线免费看| 王馨瑶露胸无遮挡在线观看| 精品人妻熟女av久视频| 热99久久久久精品小说推荐| 成年美女黄网站色视频大全免费 | 亚洲精品乱码久久久v下载方式| 色5月婷婷丁香| 国产黄色免费在线视频| 久久综合国产亚洲精品| 美女视频免费永久观看网站| 久久女婷五月综合色啪小说| 日本午夜av视频| 99九九线精品视频在线观看视频| 秋霞在线观看毛片| 国产精品人妻久久久久久| 久久久国产精品麻豆| 欧美三级亚洲精品| 精品一区二区免费观看| 丝袜美足系列| 国产免费一级a男人的天堂| 欧美xxⅹ黑人| 成人国语在线视频| 美女cb高潮喷水在线观看| 自线自在国产av| 久久av网站| 夫妻性生交免费视频一级片| 搡女人真爽免费视频火全软件| 国产成人av激情在线播放 | 天天操日日干夜夜撸| 日韩三级伦理在线观看| 国产黄色视频一区二区在线观看| 午夜免费男女啪啪视频观看| 成人黄色视频免费在线看| 人人妻人人添人人爽欧美一区卜| 国产爽快片一区二区三区| 亚洲经典国产精华液单| av又黄又爽大尺度在线免费看| 久久午夜福利片| 乱人伦中国视频| 国产在线视频一区二区| 国产极品天堂在线| 伦理电影免费视频| 下体分泌物呈黄色| 久久久久久久亚洲中文字幕| 欧美日韩综合久久久久久| 欧美人与性动交α欧美精品济南到 | 夫妻午夜视频| 简卡轻食公司| 天天影视国产精品| 久久久久久久久久人人人人人人| 国产永久视频网站| 日本vs欧美在线观看视频| 中国国产av一级| 一本久久精品| 97在线人人人人妻| 在线天堂最新版资源| 日本wwww免费看| 久久青草综合色| 久久99热6这里只有精品| 伊人久久精品亚洲午夜| 亚洲国产av影院在线观看| 亚洲精品国产av成人精品| 亚洲三级黄色毛片| 大香蕉久久成人网| 国产免费一区二区三区四区乱码| 两个人免费观看高清视频| 国产黄色免费在线视频| av卡一久久| 国产高清三级在线| 免费高清在线观看视频在线观看| 色视频在线一区二区三区| 久久久久久人妻| 亚洲人成77777在线视频| 桃花免费在线播放| 麻豆乱淫一区二区| 黑丝袜美女国产一区| 精品午夜福利在线看| 日韩亚洲欧美综合| 日本黄色日本黄色录像| 黑人高潮一二区| 老司机影院成人| 亚洲人成网站在线观看播放| 一区二区av电影网| 大香蕉97超碰在线| 最近手机中文字幕大全| 飞空精品影院首页| 亚洲欧美一区二区三区黑人 | 黄片无遮挡物在线观看| 插阴视频在线观看视频| 久久久a久久爽久久v久久| 国产精品一区二区在线观看99| 日韩中文字幕视频在线看片| 水蜜桃什么品种好| 国产白丝娇喘喷水9色精品| 桃花免费在线播放| 少妇 在线观看| 国产在线视频一区二区| 一级毛片 在线播放| 成年人午夜在线观看视频| 考比视频在线观看| 日本wwww免费看| 女性被躁到高潮视频| 国产又色又爽无遮挡免| 亚洲av男天堂| 性色av一级| 性高湖久久久久久久久免费观看| √禁漫天堂资源中文www| 国产亚洲一区二区精品| 久久人妻熟女aⅴ| 国产精品三级大全| 乱人伦中国视频| 边亲边吃奶的免费视频| 啦啦啦视频在线资源免费观看| 国产亚洲av片在线观看秒播厂| 亚洲av欧美aⅴ国产| 亚洲国产精品一区三区| 一个人免费看片子| 街头女战士在线观看网站| 精品卡一卡二卡四卡免费| 夜夜爽夜夜爽视频| 看非洲黑人一级黄片| av电影中文网址| 青春草视频在线免费观看| 欧美亚洲 丝袜 人妻 在线| 少妇猛男粗大的猛烈进出视频| 日本91视频免费播放| 久久免费观看电影| 午夜老司机福利剧场| 赤兔流量卡办理| 少妇被粗大猛烈的视频| 人妻少妇偷人精品九色| 波野结衣二区三区在线| 91久久精品国产一区二区三区| 国产精品久久久久久久久免| 国产女主播在线喷水免费视频网站| 亚洲情色 制服丝袜| 欧美成人精品欧美一级黄| 国产av精品麻豆| √禁漫天堂资源中文www| 亚洲av不卡在线观看| 免费观看在线日韩| 亚洲精品av麻豆狂野| 国产一区有黄有色的免费视频| 最新的欧美精品一区二区| 一二三四中文在线观看免费高清| 久热这里只有精品99| 99热国产这里只有精品6| 国产精品久久久久成人av| 色5月婷婷丁香| 一级毛片 在线播放| 麻豆乱淫一区二区| 一级毛片黄色毛片免费观看视频| 午夜免费鲁丝| h视频一区二区三区| 久久精品国产a三级三级三级| 亚洲成色77777| 欧美少妇被猛烈插入视频| 如何舔出高潮| 成人毛片60女人毛片免费| 免费大片18禁| 日本欧美视频一区| 一区二区av电影网| 国产精品久久久久久精品古装| 人人妻人人澡人人爽人人夜夜| 夫妻午夜视频| 国产精品麻豆人妻色哟哟久久| 久久99精品国语久久久| 亚洲国产精品成人久久小说| 最近中文字幕2019免费版| 国产精品一区二区在线不卡| 熟妇人妻不卡中文字幕| 欧美日韩成人在线一区二区| 插阴视频在线观看视频| www.av在线官网国产| 午夜激情久久久久久久| av在线app专区| 久久久午夜欧美精品| 久久午夜福利片| 日韩成人伦理影院| 国产精品免费大片| 老女人水多毛片| 黄片无遮挡物在线观看| 日本欧美国产在线视频| 久久精品人人爽人人爽视色| 校园人妻丝袜中文字幕| 欧美变态另类bdsm刘玥| 日韩大片免费观看网站| 亚洲美女搞黄在线观看| 国产精品国产三级专区第一集| 精品国产国语对白av| 国产又色又爽无遮挡免| 免费久久久久久久精品成人欧美视频 | av不卡在线播放| 十八禁网站网址无遮挡| 狂野欧美激情性xxxx在线观看| 伊人久久国产一区二区| 国产欧美日韩综合在线一区二区| 51国产日韩欧美| 国产高清不卡午夜福利| 最近中文字幕高清免费大全6| 中文字幕久久专区| 色吧在线观看| 国产爽快片一区二区三区| 日韩av在线免费看完整版不卡| 一区二区三区乱码不卡18| 久久精品国产亚洲av涩爱| 又粗又硬又长又爽又黄的视频| 成年av动漫网址| 国产精品国产三级专区第一集| 99久久精品国产国产毛片| 国产一区二区三区综合在线观看 | 黄色视频在线播放观看不卡| 熟女电影av网| 大片免费播放器 马上看| 尾随美女入室| 国产高清国产精品国产三级| 久久久精品免费免费高清| 日本vs欧美在线观看视频| 热re99久久精品国产66热6| 国产女主播在线喷水免费视频网站| 男男h啪啪无遮挡| 蜜桃在线观看..| 国产精品人妻久久久影院| 亚洲美女搞黄在线观看| 街头女战士在线观看网站| 国产片特级美女逼逼视频| 亚洲精品久久成人aⅴ小说 | 99热这里只有是精品在线观看| 人人妻人人爽人人添夜夜欢视频| 美女国产视频在线观看| 国产精品嫩草影院av在线观看| 久热这里只有精品99| 国产精品.久久久| 亚洲国产精品999| 亚洲欧美精品自产自拍| 国产国语露脸激情在线看| 亚洲一级一片aⅴ在线观看| 久久久久网色| 18+在线观看网站| 国产一区二区在线观看日韩| 午夜老司机福利剧场| 亚洲第一av免费看| 亚洲精品自拍成人| 中文字幕av电影在线播放| 精品人妻一区二区三区麻豆| 天堂中文最新版在线下载| 丝瓜视频免费看黄片| 香蕉精品网在线| 女性生殖器流出的白浆| 中国三级夫妇交换| 亚洲av免费高清在线观看| 在线观看免费视频网站a站| 亚洲精品久久久久久婷婷小说| 一区二区三区四区激情视频| 亚洲国产精品国产精品| 夜夜看夜夜爽夜夜摸| 精品午夜福利在线看| 欧美日韩亚洲高清精品| 色94色欧美一区二区| 日本与韩国留学比较| 在线观看免费高清a一片| 另类精品久久| 22中文网久久字幕| 亚洲精品久久午夜乱码| 中文字幕亚洲精品专区| .国产精品久久| 亚洲国产av新网站| 国产精品久久久久久精品电影小说| 国产成人午夜福利电影在线观看| av免费观看日本| 美女内射精品一级片tv| 日韩一区二区视频免费看| 在线观看免费日韩欧美大片 | 国产成人aa在线观看| av有码第一页| 草草在线视频免费看| av国产久精品久网站免费入址| 久久久国产欧美日韩av| 日本与韩国留学比较| 毛片一级片免费看久久久久| 久久av网站| 国产极品天堂在线| 午夜福利影视在线免费观看| 亚洲av在线观看美女高潮| 高清黄色对白视频在线免费看| 成人漫画全彩无遮挡| 另类精品久久| 欧美三级亚洲精品| 中文字幕最新亚洲高清| 欧美成人午夜免费资源| 岛国毛片在线播放| 亚洲综合色惰| 午夜激情福利司机影院| 国产男女超爽视频在线观看| 2018国产大陆天天弄谢| 飞空精品影院首页| 一级,二级,三级黄色视频| 我要看黄色一级片免费的| 欧美人与善性xxx| 精品久久国产蜜桃| 伊人亚洲综合成人网| 人人妻人人添人人爽欧美一区卜| 国产亚洲精品第一综合不卡 | 日本黄色日本黄色录像| 欧美精品一区二区大全| 一级片'在线观看视频| 免费看av在线观看网站| 久久女婷五月综合色啪小说| 少妇被粗大猛烈的视频| 精品一区二区三区视频在线| 免费看光身美女| 桃花免费在线播放| 青春草国产在线视频| 国产午夜精品一二区理论片| 国产精品无大码| 在线 av 中文字幕| 99re6热这里在线精品视频| 久久免费观看电影| a级毛片免费高清观看在线播放| 国产69精品久久久久777片| 久久久久久久久久成人| 各种免费的搞黄视频| 一本大道久久a久久精品| 亚洲人成网站在线观看播放| 91aial.com中文字幕在线观看| 久久精品久久精品一区二区三区| 亚洲国产成人一精品久久久| 国产免费福利视频在线观看| 久久综合国产亚洲精品| 精品人妻一区二区三区麻豆| 日本-黄色视频高清免费观看| 男女国产视频网站| 丁香六月天网| 中文欧美无线码| 人妻少妇偷人精品九色| 国产精品国产av在线观看| 大又大粗又爽又黄少妇毛片口| 大香蕉久久成人网| 如日韩欧美国产精品一区二区三区 | 黄片播放在线免费| 国产av一区二区精品久久| 成人国产av品久久久| a级片在线免费高清观看视频| 欧美xxⅹ黑人| 一区二区三区四区激情视频| 日韩欧美一区视频在线观看| 国产精品成人在线| 一区二区三区免费毛片| 91精品一卡2卡3卡4卡| 国产一区二区在线观看日韩| 一级毛片我不卡| 男人添女人高潮全过程视频| 天天操日日干夜夜撸| 人妻人人澡人人爽人人| 免费大片黄手机在线观看| 丰满迷人的少妇在线观看| 成年女人在线观看亚洲视频| 十分钟在线观看高清视频www| 日韩免费高清中文字幕av| 久久韩国三级中文字幕| 成人午夜精彩视频在线观看| 欧美人与性动交α欧美精品济南到 | 黄色一级大片看看| tube8黄色片| 国产爽快片一区二区三区| 亚洲精品一二三| 日韩三级伦理在线观看| 成人二区视频| 久久热精品热| 最后的刺客免费高清国语| 午夜福利视频精品| 日本黄大片高清| 精品国产一区二区久久| 午夜精品国产一区二区电影| 亚洲精品亚洲一区二区| av电影中文网址| 99re6热这里在线精品视频| 美女国产视频在线观看| 国产无遮挡羞羞视频在线观看| 国产av精品麻豆| 9色porny在线观看| 在线观看人妻少妇| 少妇丰满av| 国产精品一二三区在线看| 伊人亚洲综合成人网| 一区二区三区免费毛片| 人人妻人人爽人人添夜夜欢视频| 国产精品三级大全| 久久久a久久爽久久v久久| 男女边摸边吃奶| 91在线精品国自产拍蜜月| tube8黄色片| 久久精品国产自在天天线| 美女视频免费永久观看网站| 天天操日日干夜夜撸| 精品少妇久久久久久888优播| 免费少妇av软件| 精品久久久久久电影网| 香蕉精品网在线| 亚洲av中文av极速乱| 国产老妇伦熟女老妇高清| 国产日韩欧美在线精品| 中文字幕人妻熟人妻熟丝袜美| 国产精品人妻久久久久久| 亚洲av福利一区| 9色porny在线观看| 高清视频免费观看一区二区| 免费大片18禁| 99re6热这里在线精品视频| 青青草视频在线视频观看| 日日摸夜夜添夜夜添av毛片| 国产午夜精品久久久久久一区二区三区| 青春草国产在线视频| 如日韩欧美国产精品一区二区三区 | 成年人午夜在线观看视频| 精品国产露脸久久av麻豆| 丰满少妇做爰视频| 99视频精品全部免费 在线| 99九九在线精品视频| 国产精品久久久久久精品电影小说| 久久国产亚洲av麻豆专区| 国产一区二区三区综合在线观看 | 国产av国产精品国产| 欧美激情 高清一区二区三区| 91久久精品国产一区二区三区| 亚洲国产精品国产精品| 久久这里有精品视频免费| 99久久中文字幕三级久久日本| 国产亚洲精品第一综合不卡 | av网站免费在线观看视频| 国产av国产精品国产| 在线精品无人区一区二区三| 中文字幕最新亚洲高清| 国产精品国产av在线观看| 国产视频首页在线观看| 国产精品不卡视频一区二区| 欧美少妇被猛烈插入视频| 一级爰片在线观看| 久久精品国产自在天天线| 纯流量卡能插随身wifi吗| 国产极品天堂在线| 国产精品久久久久久精品电影小说| 欧美三级亚洲精品| 老司机影院成人| 满18在线观看网站| 午夜老司机福利剧场| 国产一区二区三区综合在线观看 | 91久久精品国产一区二区成人| 久久鲁丝午夜福利片| 亚洲情色 制服丝袜| 日日摸夜夜添夜夜添av毛片| 桃花免费在线播放| 视频中文字幕在线观看| 国产精品国产av在线观看| 91久久精品国产一区二区三区| 日韩成人伦理影院| 9色porny在线观看| 欧美国产精品一级二级三级| 丝瓜视频免费看黄片| 国产一区二区在线观看日韩| 三级国产精品片| freevideosex欧美| 蜜桃国产av成人99| 日本色播在线视频| 三级国产精品片| 制服诱惑二区| 亚洲国产精品成人久久小说| 亚洲人成网站在线播| 视频在线观看一区二区三区| 亚洲欧美日韩另类电影网站| av播播在线观看一区| 亚洲精品第二区| a级毛片在线看网站| 免费黄色在线免费观看| 成人手机av| 国产一级毛片在线| 欧美日韩精品成人综合77777| 久久久久久久国产电影| 成人国产麻豆网| 精品亚洲成国产av| 午夜久久久在线观看| 免费看不卡的av| 色视频在线一区二区三区| 97超碰精品成人国产| 高清av免费在线| 亚洲av成人精品一区久久| 日本vs欧美在线观看视频| 久久精品国产亚洲av天美| 26uuu在线亚洲综合色| av黄色大香蕉| 日韩大片免费观看网站| 国产日韩欧美在线精品| 人人妻人人澡人人看| 久久久久久久大尺度免费视频| 久久99精品国语久久久| 亚洲五月色婷婷综合| 亚洲一区二区三区欧美精品| 秋霞在线观看毛片| 男女边摸边吃奶| 日本黄色片子视频| 日韩av在线免费看完整版不卡| 国产精品99久久99久久久不卡 | 大陆偷拍与自拍| 午夜福利影视在线免费观看| 精品熟女少妇av免费看| 曰老女人黄片| 日本与韩国留学比较| 丝瓜视频免费看黄片| 新久久久久国产一级毛片| 亚洲国产精品专区欧美| 99久久人妻综合| 菩萨蛮人人尽说江南好唐韦庄| 蜜桃久久精品国产亚洲av| 久久久久久伊人网av| 在线观看一区二区三区激情| 多毛熟女@视频| 国产精品国产av在线观看| 丰满迷人的少妇在线观看| 精品少妇黑人巨大在线播放| 综合色丁香网| 欧美激情极品国产一区二区三区 | 国产精品欧美亚洲77777| 男女无遮挡免费网站观看| 又粗又硬又长又爽又黄的视频| 成年女人在线观看亚洲视频| av天堂久久9| 一本一本久久a久久精品综合妖精 国产伦在线观看视频一区 | 嘟嘟电影网在线观看| 美女主播在线视频| 丁香六月天网| 多毛熟女@视频| 午夜福利,免费看| 美女内射精品一级片tv| 久久99一区二区三区| 国产69精品久久久久777片| 3wmmmm亚洲av在线观看| 老熟女久久久| 成人国产av品久久久| 国产一区有黄有色的免费视频| 桃花免费在线播放| 欧美97在线视频| 国产高清国产精品国产三级| 不卡视频在线观看欧美| 美女内射精品一级片tv| 永久网站在线| 三级国产精品片| 99国产精品免费福利视频| 桃花免费在线播放| 精品人妻一区二区三区麻豆| 欧美国产精品一级二级三级| 日日摸夜夜添夜夜添av毛片| 久久97久久精品| 你懂的网址亚洲精品在线观看| 亚洲国产精品成人久久小说| 久久99精品国语久久久| 久久久久久久大尺度免费视频| 国产欧美日韩综合在线一区二区| 在线观看一区二区三区激情| 99九九线精品视频在线观看视频| 成人亚洲精品一区在线观看| 久久久久久久久大av| 免费黄网站久久成人精品| 久久人人爽人人爽人人片va| 人妻制服诱惑在线中文字幕| 2021少妇久久久久久久久久久| 久久久亚洲精品成人影院| 久久久久久人妻| 久久久国产一区二区| 一个人看视频在线观看www免费|