• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    CurveNet: Curvature-Based Multitask Learning Deep Networks for 3D Object Recognition

    2021-06-18 03:36:46MuzahidWanggenWanFerdousSohelLianyaoWuandLiHou
    IEEE/CAA Journal of Automatica Sinica 2021年6期

    A. A. M. Muzahid, Wanggen Wan,, Ferdous Sohel,,Lianyao Wu, and Li Hou

    Abstract—In computer vision fields, 3D object recognition is one of the most important tasks for many real-world applications.Three-dimensional convolutional neural networks (CNNs) have demonstrated their advantages in 3D object recognition. In this paper, we propose to use the principal curvature directions of 3D objects (using a CAD model) to represent the geometric features as inputs for the 3D CNN. Our framework, namely CurveNet,learns perceptually relevant salient features and predicts object class labels. Curvature directions incorporate complex surface information of a 3D object, which helps our framework to produce more precise and discriminative features for object recognition. Multitask learning is inspired by sharing features between two related tasks, where we consider pose classification as an auxiliary task to enable our CurveNet to better generalize object label classification. Experimental results show that our proposed framework using curvature vectors performs better than voxels as an input for 3D object classification. We further improved the performance of CurveNet by combining two networks with both curvature direction and voxels of a 3D object as the inputs. A Cross-Stitch module was adopted to learn effective shared features across multiple representations. We evaluated our methods using three publicly available datasets and achieved competitive performance in the 3D object recognition task.

    I. INTRODUCTION

    IN the field of computer vision, 2D image analysis using deep learning (DL) methods has already achieved remarkable progress and outperforms human vision in many cases (e.g., image classification and human face analysis) [1],[2]. However, understanding three-dimensional objects is still an open research problem of modern computer vision research. A real object in three-dimensional space provides more detailed information. With the availability of low-cost 3D acquisition devices, it is easier to capture 3D objects. The rise of public repositories of 3D models has drawn attention to computer vision research such as 3D object recognition,reconstruction, semantic segmentation, and retrieval [3].

    Convolutional neural network (CNN)-based 3D object recognition systems have advanced considerably [4], but 3D CNNs have been not as successful in identifying 3D objects as 2D CNNs, especially in object label prediction. There are several reasons behind this, e.g., selecting the input features of a 3D object [5] is critical due to its complex geometrical structures, comparatively smaller training databases, and the high computational cost required by 3D CNNs. The earliest volumetric deep learning approach is 3D ShapeNets [6],which deals with three tasks including 3D object classification. Recently, several approaches have been published that solve 3D object recognition tasks using deep CNNs; voxels [6]-[9], point clouds [10]-[12], and 2D multiview [13]-[15] are the most widely used representations of CNNs for 3D object recognition [4]. Two-dimensional multiview-based approaches that use 2D CNNs achieve high performance. This is because existing 2D-based CNNs can be directly used for 3D object recognition, and they require fewer computational resources. However, multiview representations have some technical issues; for example, choosing the number of views to capture the information of the entire 3D object is still an open issue. In addition, the projection of 3D data to the 2D domain discards intrinsic features (e.g., geometric,structural, and orientational information) of a 3D object.

    In 3D shape analysis, the three-dimensional representation is the only way to preserve the intrinsic features of a 3D object; therefore, new features of a 3D object and advanced neural networks need to be explored to improve 3D vision tasks [8]. AI-based computer vision systems are developed with advanced machine learning (ML) algorithms (e.g., deep learning). In addition, the object classification accuracy of a CNN is also highly influenced by the input features [4].

    In this study, we incorporated curvature directions as input features of a 3D object into our novel 3D CNN to identify object labels. Principal curvature directions of 3D objects are considered as perceptually relevant properties of the human visual system (HVS) that are widely used in 3D mesh visual quality evaluation [16], [17]. Considering perceptual features in the HVS, curvature maps represent the salient structural features of a 3D object [17]. The curvature map of a typical 3D CAD model is shown in Figs. 1(d) and 1(e). The plot of curvature vectors looks similar to a sketch of the original 3D model. The idea is that if a typical neural network model can recognize an object from the boundary information (from voxels) of a 3D object, then it can also recognize it from its sketch.

    Fig. 1. The comparison of different 3D representations (a) original 3D Bimba CAD model; (b) voxel representation; (c) point-cloud representation;(d) curvature direction map using maximum values [17]; and (e) curvature direction map using minimum values [17].

    Inspired by this, we propose CurveNet for 3D object classification, which is a novel volumetric CNN that inputs curvature points directly as low-level input features. In CNNs,multitask [18], [19] and multiscale [20], [21] learning-based approaches provide additional benefits that improve the performance of the network for object classification. In multitask learning, an auxiliary task is considered to optimize the loss and improve the performance of the main task. We added pose recognition as an auxiliary task to improve the performance of CurveNet for the object category-level classification. In general, the classification accuracy using curvature points should not be worse than that of voxel input as they provide more information (discriminative surface feature), especially for the curve surface or complex structured 3D model. Our experimental results demonstrate that curvature-based input features outperform the object boundary with voxel representations. A new inception module was introduced to generate a high-level global feature from the low-level surface input feature. Our proposed CurveNet benefits from multiscale learning from this new inception module. To improve the network performance further, we fused the two networks, which take curvature vector points and voxels as inputs. Data augmentation was applied to increase the number of training samples, which helped to reduce the risk of overfitting and improve the performance of the network.

    The key contributions in this paper are as follows: i) We propose to use curvature information of a 3D CAD model as an input for the 3D classifier, CurveNet. To the best of our knowledge, this is the first time that a 3D CNN has directly taken curvature vectors as inputs and outperformed boundary representation from voxels. ii) We introduced a new inception block consisting of three conv layers, and one max pooling layer, which is an efficient alternative to the conventional multiscale learning strategy. This improves the learning capability of CurveNet and generates very high-level global features with fewer parameters. iii) We extended our technique by combining multiple inputs (boundary information from voxels and curvature directions) and experimented with the influence of soft and hard parameter sharing to improve the performance of 3D object classification.

    The rest of the paper is organized as follows: The related work on 3D object recognition based on neural networks is reviewed in Section II. Section III describes the proposed CurveNet architecture including the feature extraction method, network architecture, and network fusion techniques.Network training and the experimental results and analysis are presented in Section IV. Finally, we draw a conclusion in Section V.

    II. RELATED WORKS

    The most common representations of 3D objects used for 3D object recognition are multiview, point clouds, volumetric,hypergraph, and 3D descriptors.

    1) Multiview:One of the earliest approaches to 3D object recognition was presented in the context of RGB-D images[22], [23]. This feature learning approach is similar to image classification, where an additional input is used for the depth information. The multiview convolutional neural network(MVCNN) [13] model is a pioneering multiview-based CNN model, in which a 3D object can be rendered to 2D images from 12 to 80 viewpoints. A pre-trained model from ImageNet[20] was used to train the network model on the ModelNet40 dataset [24] for the 3D classification task. RotationNet [25]was proposed to jointly predict the object label and pose using multiview images of a 3D object. To use the correlation information between the multiview image of a 3D object, Maet al.proposed [26] a novel multiview-based network for classification and retrieval tasks. However, multiview representation discards the geometric properties, and it is hard to determine the number of projections required to capture the entire surface of a 3D object.

    2) Point Cloud:The point cloud representations are popular for their ease of 3D shape representation and their direct usage in CNNs as inputs. PointNet [12] is one of the earliest approaches proposed to address the problems of 3D object classification and segmentation tasks. PointNet directly inputs the pointset and outputs the object class label. However,PointNet does not capture the local structure induced by the pointwise feature extraction method. This problem is addressed in PointNet++ [11] by processing a set of points sampled in a hierarchical order. Ben-Shabatet al.[27]proposed a hybrid point cloud representation, namely 3D modified fisher vectors (3DmFV), which efficiently combines discrete point structures with a continuous generalization of fisher vectors for 3D point cloud classification and part segmentation. Point2Sequence [10] was proposed by Liuet al.as a means of capturing fine-grained contextual information and learning point cloud features by aggregating multi-scale areas of each local region of 3D objects with attention. The relation-shape convolutional neural network (RS-CNN) [28]was introduced to learn contextual shape-aware information for point cloud classification and segmentation problems.However, point clouds are somewhat different from rasterized data (voxels or pixels), which are formed in a spatial irregular fashion, but the main benefit of their coordinates can be used straightforwardly in CNNs as inputs. Due to the advancement of 3D scanning tools, it is very easy and fast to develop a point cloud model, but the model fails to preserve the surface information of the object in the real environment because of the unordered point sampling.

    3) Volumetric:Volumetric representation is a very powerful representation that can describe the full geometry of a 3D object. It holds the intrinsic feature of viewpoint information.Voxel and octree are the most commonly used representation methods of volumetric data for 3D object recognition tasks[3]. Wuet al.proposed a deep learning model for volumetric shape analysis including recognition, namely 3DShapeNets[6], the pioneering and earliest approach to 3D object recognition with volumetric data. The following outlines several fruitful volumetric approaches proposed using 3D voxel grid representation [5], [7], [12], [25]. Recently, octreebased volumetric CNNs [30]-[32] significantly improved the performance of 3D object recognition, consuming less memory compared to voxel-based methods. In addition, there are several new approaches introduced to 3D volumetric CNNs such as multitask learning [33], including sub-volume supervision [9], [14], [32], and network fusion methods with multi-representation input, which have gained noticeable progress [8], [14], [34] in recent years. In general, volumetric CNNs require higher computational cost, which may be challenging for real-time applications. VoxNet and LightNet addressed this issue and introduced volumetric networks with the smallest parameters of about 0.92 M and 0.30 M,respectively. Brocket al.proposed the voxel-based Voxception-ResNet (VRN) [35] model that learned latent space using a variational autoencoder (VAE). The VRN ensemble has achieved the best classification accuracy until now on the ModelNet public dataset [24]. Ghadaiet al.[36]introduced a multi-level 3D CNN that learned multi-scale features by utilizing a coarse voxel grid representation.

    4) Others:Lucianoet al.introduced a deep similarity network fusion framework, namely DeepSNF [37], for 3D object classification, where features from the geometric shape descriptor are fed into a graph convolutional neural network[38]. Recently, several methods were introduced using graph CNNs [38], as these CNNs can process both sparse and unordered data. Among them, dynamic graph [39],hypergraph [40], and multi-hypergraph [41] stereographicbased methods achieve significant improvement in 3D object recognition tasks. These aforementioned methods follow supervised learning strategies; in addition, there are several unsupervised methods [42]-[44] introduced for 3D object recognition. However, the performance of unsupervised approaches is generally worse than that of supervised methods.

    III. PROPOSED METHOD

    We consider a volumetric CNN in this paper. In volumetric data, voxels are commonly used to represent a 3D object.However, to improve the performance and reduce the computational cost of volumetric methods, several approaches have been introduced, such as octree [26], [28], combined voxel with curvature map [45], and normal vector [8], used as an input for a 3D CNN. There is still a large performance gap between 3D and 2D CNNs as 2D-based methods benefit from techniques from numerous existing works in image processing and perform better in 3D object recognition tasks. To reduce this gap, we propose a novel 3D CNN network, called CurveNet, which uses curvature points as inputs. The main reason for using a 3D CNN is to extract rich information from spatial structures and learn the fine details from the non-plane surface at various orientations. This section presents input data preparation, CurveNet architecture, and its learning strategy in detail.

    A. Input Data Preparation

    Two popular ways of representing a 3D object with a CNN are multiview and volumetric. Multiview representation is a 2D image of a 3D object captured from multiple viewpoints that discard the structural information. On the other hand,volumetric representation preserves the intrinsic features with viewpoint information, which is widely used as an input of 3D CNNs.

    We extracted the curvature direction map of a 3D mesh (3D CAD model). The irregular curvature point vectors are formed into a 3D data tensor, which is encoded further to regular points that are input into the 3D CNN. To calculate curvature direction in 2D when applied to a circle of radiusr(Fig. 2), let us define two pointse1ande2on the circle, and their normals aren1andn2respectively from the circle’s center. The curvature of the circle can be calculated as

    The curvature calculation from (1) can be extended on arbitrary 3D meshes by projecting the vectorn1-n2onto the curvature edge as

    Fig. 2. Circular curvature computation in 2D.

    Fig. 3. Geometric elements used for curvature computation for a 3D mesh.

    To extract these curvature vectors (min, max), we used an open-source tool, mepp1https://github.com/MEPP-team/MEPP(3D mesh processing platform), with modified TPDM metrics [17] on Ubuntu. The curvature vectors (3D vectors for three dimensions) are calculated in an iterative manner and stored as unstructured points (3D tensors) on the local drive. To encode these points, we cut 3D tensors of each 3D model with 5k points, and distributed them uniformly by Poisson disk sampling using Meshlab [47] and transformed them into a 3D grid [7] as the input for the CNN.Fig. 1 shows a representation of a 3D CAD model from voxels, point clouds, and curvature points.

    B. CurveNet Network Architecture

    Our proposed CurveNet is a 3D deep convolutional neural network, consisting of conv and FC layers. Fig. 4 depicts the general block diagram of CurveNet. The network is designed in the form of a multitask learning strategy. In multitask learning, at least two or more tasks are assigned [18]. Between them, one is the main task, and the rest of the tasks are auxiliary tasks. Auxiliary tasks are used to improve the performance of the main task. In this work, object label prediction is the main task, and pose prediction is the auxiliary task of CurveNet. CurveNet directly inputs the binary of the 3D data tensor by its input layer, which has a size of 30×30×30. A new inception block was equipped to generate discriminative features using a multiscale training approach.We used three identical inception blocks. Each inception block consisted of three conv layers and one max-pooling layer. Convolution parameters are represented as (C,S) ×N,whereCis the convolution kernel,Sis stride, andNrefers to the number of output feature maps. The output of each conv layer is passed to a sequence of a batch normalization layer, a scale layer, and a leaky rectified linear unit (Leaky ReLU)layer with a negative slope of 0.1.

    We used a typical form of a max-pooling layer. We applied a non-overlapping pooling function using kernel size 2 with a stride of 2, so the resolution of the feature map decreased by a factor of 2. The output of conv-2 and conv-3 is the same with the number of output feature maps that are concatenated to be input into the conv layer of the next inception block.However, conv-2 and conv-3 input features from the output of conv-1 and max-pooling layers, respectively, where the dimensions of two input features are (d/2-1), and (d/2),respectively. These multiscale features are produced by the inception block and sent to the next block as an input, which generates more distinguishable features to recognize object class labels. The number of learned feature maps by Inception-1 is 112, and this is increased twice (224, and 448)by the following Inception blocks. CurveNet uses four FC layers at the top, where the last two FCs, FC-3 and FC-4, are used to get the prediction score of object label and pose,respectively. FC-3 and FC-4 both take the output of FC-2 as an input. We used a softmax classifier to compute the accuracy and loss [48] of CurveNet for object label classification tasks.

    C. Parameter Sharing by a Combination of Two Inputs

    In this work, we employed network fusion strategies on CurveNet using both curvature directions and surface boundary representation from voxels as inputs. To experience the influence of parameter sharing, two CurveNets were fused to improve the classification performance, which takes boundary information from voxels [7] and curvature information as inputs. We trained our CurveNet using three different fusion strategies, namely no sharing, hard parameter sharing, and soft parameter sharing. All fusion networks are illustrated in Fig. 5. Training samples were chosen from the same object with similar volume resolutions for two inputs.The aim was to train a fusion of CurveNets to improve the classification accuracy using multiple inputs and find an effective feature sharing strategy.

    Fig. 4. CurveNet framework. A new Inception module is proposed. Inceptions 1, 2, and 3 are identical but differ with the number of input and output features and parameters (C1 = 3, C2 = 1, C3 = 2, S1 = 2, S2 = 1, and m = 2). The total number of training parameters is 3.1 M.

    Fig. 5. Parameter sharing using network fusion methods with two CurveNets (a) no sharing; (b) hard parameter sharing; and (c) soft parameter sharing.

    The first fusion pipeline is shown in Fig. 5(a). This pipeline is trained with no sharing strategy, and all learned features are concatenated before classification. The classification decision comes from the linear combination of FC outputs. The combination of learned features by two networks is fused before classifiers (topmost FCs). Then, the high-label global features are shared to classify the object into two categories.Our goal was to identify the correct object label (main task).We designed the networks to predict the object pose as an auxiliary task to improve the network performance using orientation information for our desired task [18].

    In the second fusion pipeline, we follow the hard parameter sharing (see Fig. 5(b)) where one or more conv and FC layers can share their learned features. We shared parameters between Inception-2 and Inception-3 layers. We found that sharing all layers including FCs did not help at all to increase the performance but consumed 50 percent more training time.In multitask learning, some layers should be invariant for auxiliary tasks [18], and we found that two separate branches of FCs provided better results than shared FCs.

    We implemented soft-parameter sharing by adopting Cross-Stitch networks [49], as illustrated in Fig. 5(c), where it was determined whether distinct learned features were shared or learned. The weighted parameters of two networks are sharable and trainable. The decision of sharing to the following layers is mediated by α, a matrix of a linear combination of activation maps. The mediated parameter decides whether the features are shared or learned.Accordingly, the network distributes task-related features to improve the prediction scores of the desired task. Finally, all learned features inside the network are forwarded to classifier layers through the final mediator block, which has the ability to select suitable and important features for the specific task(label or pose recognition). In both cases, the final output layer (softmax function) of two networks is averaged, and the largest prediction indicates the target class. All our fusion networks are followed by a layer-based parameter sharing strategy.

    IV. EXPERIMENTS AND ANALYSIS

    A. Training and Evaluation Setup

    We evaluated the classification performance of our proposed CurveNet on two different types of 3D objects,specifically 3D CAD models and LiDAR point clouds using Princeton ModelNet [24] and Sydney Urban objects datasets,respectively. Our CurveNet was implemented on the TensorFlow deep learning framework, which was installed on a Xeon-x5650 CPU with a Teslak20c GPU enabled by cuDNN8.0.

    We trained our CurveNet end-to-end in a supervised manner by stochastic gradient descent (SGB) [50] with the momentum parameter set to 0.9. The batch size was set to 32. Initial learning was set to 0.01 and decreased by a factor of 10. We added dropout regularization after each output layer with a ratio of 0.6. We computed the multiclass loss function using multinomial cross-entropy losses [51]. Our CurveNet took approximately 26 hours to be trained for 90 epochs. We applied voting strategy during testing where multiple rotations of a sample 3D object were fed to the network, and outputs were averaged to get the final prediction of the object class label.

    B. Data Alignment and Augmentation

    Modelnet40 and ModelNet10 are the two main datasets widely used to evaluate classification performance for 3D objects. However, the 3D models in ModelNet10 are aligned perfectly; in contrast, more than 1000 models in ModelNet40 are not aligned. We first sorted the classes of non-aligned objects and prepared a reference set for each class to be aligned. We adopted an unsupervised generation of viewpoint[52] methods to align objects according to class, and labeling of object pose was assigned iteratively by running three different subsets using 90-degree rotations. Some rotationally invariant objects (e.g., samples in flower pots and bottles)were forcibly assigned to class 1 as they did not provide any significant information in different poses.

    The visual attribute of a 3D object is influenced by the object pose and the number of samples per class. We extended the dataset samples by rotating each 3D model around the horizontal direction along thez-axis. We prepared two augmented subsets of the original datasets. We generated 12 and 24 copies of each 3D model by rotating by intervals of 30 and 15 degrees for two subsets, respectively. This augmentation procedure has been applied for both ModelNet40 and ModelNet10 datasets. This augmentation method helps to increase the number of training samples of each class (both pose and category) and helps to improve the performance of the network.

    C. Classification Performance on ModelNet Datasets

    The ModelNet datasets were built especially to evaluate 3D object classifiers by Changet al.[24] in 2015. ModelNet40 is a large-scale dataset containing 12311 3D CAD models that are classified into 40 categories. ModelNet10 is a small dataset that is a subset of ModelNet40, containing 4899 3D CAD models in 10 categories. The datasets come with their own training and testing splits. For consistency, we used the original splits from the datasets in our experiments. During the test, we sent all augmented models of a test sample to the network in one batch, and the object category was decided by averaging the activations of the output layer.

    1) Classification Results:We evaluated our model using both surface boundary information from voxels and curvature information. The classification accuracy (average class) is presented in Table I. We found that the normalized maximum curvature direction vector showed superiority over voxels.Our proposed method improves the classification accuracy by 1.2% compared to surface boundary information using the same network. It also outperforms 3DShapeNet [6] and VoxNet [7], which are state-of-the-art volumetric methods for 3D object classification. The improvements are mainly the result of two factors: curvature information as an input feature and auxiliary learning by adding pose classification. Curvature provides considerable depth and perceptual features to the CNN, and auxiliary pose learning helps to improve object classification accuracy, which is the main task of our network.In addition, the voting method is another common approach to improving the performance of the classifier. Table II shows the influence of pose learning and voting methods on improving the classification accuracy using our CurveNet. We used two augmented datasets from ModelNet40 using 12 and 24 copies of each sample. Experimental results show that classification accuracy is improved by 3.8% using 12-view and 3.2% using 24-view while using both the pose learning and voting approach.

    The best performance was achieved using 24-view, with an average classification accuracy of 90.2% on ModelNet40.

    From these experiments, we found that the number of training samples has an effect on the classification performance, i.e., the class with a large number of training samples has high accuracy. However, high interclass similarity between two classes can lead to a slight deviation in this trend, e.g., Chair vs. Stool and Dresser vs. Nightstand (seeFig. 6). Similar observations have also been reported in earlier works [30], [32]. When we visualize the confusion matrix of ModelNet40 in Fig. 7, it illustrates that this ambiguity only occurs if the target object belongs to some similar categories of objects, e.g., Dresser, Flower Pot, Sink, and Vase, which are examples of most confused objects in the database as they have close visual properties in instances where our network misclassified some of them.

    TABLE I COMPARISON OF CLASSIFICATION ACCURACY OF OUR CURVENET AND STATE-OF-THE-ART-METRICS ON MODELNET40

    TABLE II THE INFLUENCE OF POSE CLASSIFICATION AND VOTING METHODS ON CLASSIFYING OBJECT LABELS ON MODELNET40

    Fig. 6. Classification accuracy per class on ModelNet40 dataset.

    2) Parameter Sharing:To improve the classification performance further, both voxel and curvature features were fused using our proposed volumetric model. We applied three methods of parameter sharing between two networks, which were no sharing, hard sharing, and soft sharing, as shown in Fig. 5. Both networks (for voxel and curvature) were pretrained independently to reduce the training time.

    In the case of no sharing, generated features from different layers were concatenated before the last FC (classifier) layer.Moreover, for hard sharing, Inception-1 and Inception-2 layers were shared, then shared features were sent to FC layers consecutively. For channel-wise attention, the generated features across all layers were shared using a crossstitch algorithm (see [55] for details).

    Overall, the classification results of fusion networks were improved compared to a single network. The results are presented in Table III. We find that soft sharing has a great impact on achieving good results. It achieved average class and sample accuracy of 90.7% and 93.5%, respectively.Moreover, hard sharing performs slightly better than no sharing, as the average class and sample accuracy increased by 0.2% and 0.6%, respectively.

    Fig. 7. The confusion matrix of the ModelNet40 dataset shows the similarity between a pair of 3D objects; lower diagonal values indicate a higher degree of similarity between the objects.

    The experimental result indicates that the network with soft sharing benefited from mediated features (see Section III-C),by which it was decided whether the parameters needed to be shared or learned. On the other hand, sharing all parameters by hard sharing does not help to improve the classification result significantly; rather, using too many combinations of parameter coefficients may produce a more confusing scorethat may lead the network to a locally optimal solution.

    TABLE III PERFORMANCE EVALUATION OF CURVENET WITH SHARING FEATURES BY COMBINING CURVATURE AND BOUNDARY INFORMATION FROM VOXELS ON MODELNET40 DATABASE

    3) Comparison With Other Methods:The classification accuracy of our model is compared with state-of-the-art methods in Table IV. We found that Multiview and pointcloud methods still dominated in the 3D object classification task. In addition, our method achieved the highest accuracy among volumetric methods but was inferior to the VRN ensemble [35]. Our CurveNet has 3.1 M parameters, whereas fusion contains about 6.2 M. In comparison to the 90 M parameters of the VRN ensemble, our network is lighter, and the performance is comparable to other methods. In addition,our network also performed well on a low scale ModelNet10 database and achieved an average classification accuracy of 94.2%, which is also comparable to other methods.Theoretically, deep networks should perform better on large datasets as they need more training samples. However,samples in ModelNet10 are built with some selected classes with clean objects from ModelNet40. The confusion matrix for ModelNet10 is shown in Fig. 8, where higher diagonal values exhibit fewer similarities with other objects. In comparison to ModelNet40, all classes in ModelNet10 obtained a higher confusion score, which implies a lower similarity between the objects, resulting in better classification accuracy achieved over the ModelNet40 dataset.

    D. Sydney Urban Objects Dataset

    The Sydney Urban Objects dataset2http://www.acfr.usyd.edu.au/papers/SydneyUrbanObjectsDataset.shtml.consists of labeled Velodyne LiDAR scans of 631 urban road objects (point cloud samples) in 26 distinct categories including vehicles,pedestrians, signs, and trees. As it provides single 3D point cloud samples scanned on a common scenario of urban road,we considered this dataset to be of special interest in terms of object classification and observing the performance of our CurveNet. Because of unordered point-samples, we converted the points to volumetric representation [7] by putting all points around the objects in a bounding box and encoding them to a 303occupancy grid voxel. We empirically set the voxel size to 0.2 m. Data augmentation and test voting wereapplied accordingly with 24 rotations around thez-axis.

    TABLE IV THE COMPARISON OF CLASSIFICATION ACCURACY ACHIEVED BY OUR CURVENET AND OTHER METHODS ON MODELNET40 AND MODELNET10 DATASETS

    Fig. 8. Confusion matrix of ModelNet40 dataset.

    Classification Results:To compare the classification result with state-of-the-art methods, we followed the original dataset’s protocol [56]. The dataset was folded to four standard training and testing splits for a subset of 588 samples in 14 classes. This dataset was smaller than ModelNet;therefore, we still kept a drop-out connection of 0.6 to mitigate the overfitting during training. We report the classification performance by the averageF1score. We present the classification performance of different methods in Table V. Our model obtained an averageF1of 79.3%. The classification performance of CurveNet is significantly better than those of volumetric methods such as VoxNet [7], ORION[33], and NormalNet [8] but is slightly worse than that of LightNet [9]. This improvement implies that the classification performance on a low scale dataset is also influenced by combined multiscale features using our novel inception module and auxiliary learning on orientation information.

    TABLE V CLASSIFICATION ACCURACY OF DIFFERENT VOLUMETRIC METHODS ON SYDNEY URBAN OBJECTS DATASET

    V. CONCLUSION

    In this paper, we introduce CurveNet, which uses a curvature direction vector with a 3D CNN for 3D object recognition. Our framework takes into account multitask(label and pose) and multiscale learning (in an inception module) strategies. We applied network fusion and data augmentation methods to improve the recognition rate. Using the relevant perceptual features, our proposed CurveNet performed better than existing voxel representation models.The results are competitive with state-of-the-art point-cloudbased and multiview-based methods. In contrast, our model only works for 3D triangle mesh (3D CAD models). A weakness of our model is that it is not suitable for other 3D data representations (e.g., point clouds).

    日日撸夜夜添| 看十八女毛片水多多多| 国产 一区 欧美 日韩| 九九爱精品视频在线观看| 久久人妻av系列| 国产精品人妻久久久影院| 国产探花在线观看一区二区| 国产精品不卡视频一区二区| av天堂中文字幕网| 变态另类成人亚洲欧美熟女| 91午夜精品亚洲一区二区三区| 欧美bdsm另类| 亚洲精品粉嫩美女一区| 哪里可以看免费的av片| av在线播放精品| 男插女下体视频免费在线播放| 亚洲人与动物交配视频| 久久久久性生活片| 欧美另类亚洲清纯唯美| 精品久久久噜噜| 国产亚洲精品av在线| 在线观看66精品国产| 日本黄大片高清| 久久久久久久久久久免费av| 欧美日韩一区二区视频在线观看视频在线 | www.av在线官网国产| 啦啦啦韩国在线观看视频| 欧美另类亚洲清纯唯美| 亚洲图色成人| 久久精品国产鲁丝片午夜精品| av黄色大香蕉| 国产精品久久电影中文字幕| 久久久a久久爽久久v久久| 成人欧美大片| 欧美高清成人免费视频www| 九草在线视频观看| 身体一侧抽搐| 村上凉子中文字幕在线| 免费av毛片视频| 亚洲无线在线观看| 日日干狠狠操夜夜爽| 国产精华一区二区三区| 啦啦啦啦在线视频资源| 国产日韩欧美在线精品| 天天一区二区日本电影三级| 国产精品一区二区三区四区免费观看| 日韩在线高清观看一区二区三区| 九九爱精品视频在线观看| 在线播放国产精品三级| 国产探花在线观看一区二区| 伊人久久精品亚洲午夜| 婷婷精品国产亚洲av| 午夜福利在线观看免费完整高清在 | 国产伦在线观看视频一区| 色5月婷婷丁香| 久久鲁丝午夜福利片| 波野结衣二区三区在线| 国产爱豆传媒在线观看| 亚洲精品自拍成人| 国内揄拍国产精品人妻在线| 亚洲av免费在线观看| 噜噜噜噜噜久久久久久91| 欧美在线一区亚洲| 国产老妇伦熟女老妇高清| 蜜桃亚洲精品一区二区三区| 美女高潮的动态| 国产精品免费一区二区三区在线| 欧美又色又爽又黄视频| 免费看a级黄色片| 欧美日韩在线观看h| 黄片无遮挡物在线观看| 亚洲精品成人久久久久久| 丰满人妻一区二区三区视频av| 国产视频首页在线观看| avwww免费| av国产免费在线观看| 我的女老师完整版在线观看| 18禁在线播放成人免费| 春色校园在线视频观看| 干丝袜人妻中文字幕| 老女人水多毛片| 日日摸夜夜添夜夜添av毛片| 国产毛片a区久久久久| 久久亚洲国产成人精品v| 国产精品美女特级片免费视频播放器| 国产又黄又爽又无遮挡在线| 乱码一卡2卡4卡精品| 日韩成人伦理影院| 午夜福利高清视频| 成人国产麻豆网| 国产精品人妻久久久影院| 深爱激情五月婷婷| 久久午夜亚洲精品久久| 亚洲国产精品久久男人天堂| 老师上课跳d突然被开到最大视频| 欧美最黄视频在线播放免费| 国产精品国产高清国产av| 欧美日韩综合久久久久久| 又粗又爽又猛毛片免费看| 3wmmmm亚洲av在线观看| 丰满乱子伦码专区| 国产成人91sexporn| 波多野结衣高清作品| 丰满的人妻完整版| 色综合站精品国产| 亚洲高清免费不卡视频| 精品无人区乱码1区二区| 蜜桃久久精品国产亚洲av| 国产精品野战在线观看| 国产精品综合久久久久久久免费| a级毛片免费高清观看在线播放| 99热全是精品| 成人永久免费在线观看视频| 婷婷色综合大香蕉| 毛片一级片免费看久久久久| 乱码一卡2卡4卡精品| 亚洲成人久久性| 国产一区二区亚洲精品在线观看| 国产精品一二三区在线看| 国产成人freesex在线| 最新中文字幕久久久久| 日韩欧美精品免费久久| 久久久精品大字幕| 国产三级中文精品| 69人妻影院| 亚洲不卡免费看| 美女脱内裤让男人舔精品视频 | 亚洲欧美日韩高清专用| 少妇的逼水好多| 91aial.com中文字幕在线观看| 91在线精品国自产拍蜜月| 欧洲精品卡2卡3卡4卡5卡区| 欧美日韩一区二区视频在线观看视频在线 | 能在线免费看毛片的网站| 欧美一级a爱片免费观看看| 日韩一本色道免费dvd| 青青草视频在线视频观看| 国产精品久久电影中文字幕| 亚洲第一区二区三区不卡| 国产黄a三级三级三级人| 麻豆乱淫一区二区| 亚洲国产精品合色在线| 亚洲av男天堂| 国产精品一区www在线观看| 在线播放无遮挡| 26uuu在线亚洲综合色| 在线观看午夜福利视频| 美女被艹到高潮喷水动态| 91aial.com中文字幕在线观看| 99久久无色码亚洲精品果冻| 国产大屁股一区二区在线视频| 国产精品综合久久久久久久免费| 观看美女的网站| 国内久久婷婷六月综合欲色啪| 亚洲欧美精品自产自拍| 丰满乱子伦码专区| 欧美极品一区二区三区四区| 国产成人精品久久久久久| 天堂√8在线中文| 人妻系列 视频| 免费无遮挡裸体视频| 黑人高潮一二区| 国产真实乱freesex| 亚洲无线观看免费| 亚洲欧美成人综合另类久久久 | 99热精品在线国产| 亚洲乱码一区二区免费版| 亚洲高清免费不卡视频| 国产精品野战在线观看| 亚洲精品成人久久久久久| 美女cb高潮喷水在线观看| 亚洲av中文字字幕乱码综合| 精品久久久久久久久亚洲| 亚洲最大成人手机在线| 欧美日本视频| 亚洲欧美成人综合另类久久久 | 亚洲人成网站在线观看播放| 国产一区二区在线观看日韩| 国产亚洲精品久久久com| kizo精华| 久久精品国产自在天天线| 国产av不卡久久| 国产真实乱freesex| 国产在线精品亚洲第一网站| 三级经典国产精品| 亚洲欧美成人综合另类久久久 | 在线免费观看不下载黄p国产| 国产午夜精品论理片| 久久精品久久久久久噜噜老黄 | 一级黄片播放器| 天堂av国产一区二区熟女人妻| 人体艺术视频欧美日本| 网址你懂的国产日韩在线| 午夜a级毛片| 亚洲婷婷狠狠爱综合网| 一级av片app| 美女xxoo啪啪120秒动态图| 午夜精品在线福利| 十八禁国产超污无遮挡网站| 国产精品.久久久| 欧美bdsm另类| 久久综合国产亚洲精品| 人妻少妇偷人精品九色| 国产精品麻豆人妻色哟哟久久 | 久久草成人影院| kizo精华| 只有这里有精品99| 丰满乱子伦码专区| 午夜精品国产一区二区电影 | 日韩高清综合在线| 亚洲在线自拍视频| 午夜福利高清视频| 国产精品精品国产色婷婷| 一本一本综合久久| 久久久午夜欧美精品| 国产成人精品婷婷| 最近的中文字幕免费完整| av专区在线播放| 色哟哟哟哟哟哟| 91久久精品国产一区二区三区| 99热全是精品| 国产成人一区二区在线| 欧美一区二区精品小视频在线| 国产成年人精品一区二区| 能在线免费看毛片的网站| 中出人妻视频一区二区| 免费观看精品视频网站| 日本黄大片高清| a级毛片免费高清观看在线播放| 亚洲国产色片| 亚洲国产精品国产精品| 国产日韩欧美在线精品| 久久久久国产网址| 全区人妻精品视频| www日本黄色视频网| 丝袜美腿在线中文| 精品久久久久久成人av| 国产成人freesex在线| 欧美日韩在线观看h| 嫩草影院新地址| 如何舔出高潮| 中国美白少妇内射xxxbb| 国产极品天堂在线| 淫秽高清视频在线观看| 欧美3d第一页| av福利片在线观看| 乱人视频在线观看| 国产在线精品亚洲第一网站| eeuss影院久久| 乱系列少妇在线播放| 国产精品.久久久| 亚洲最大成人av| 国产又黄又爽又无遮挡在线| 欧美日韩精品成人综合77777| 男人的好看免费观看在线视频| 国产亚洲精品av在线| 亚洲欧美成人综合另类久久久 | а√天堂www在线а√下载| 一进一出抽搐动态| 久久久久免费精品人妻一区二区| 国产午夜精品久久久久久一区二区三区| 男女边吃奶边做爰视频| 国内精品宾馆在线| 赤兔流量卡办理| 久久国产乱子免费精品| 欧美zozozo另类| 亚洲国产精品久久男人天堂| h日本视频在线播放| 国产精品一及| 午夜爱爱视频在线播放| 中文资源天堂在线| avwww免费| 午夜福利视频1000在线观看| 国产精品野战在线观看| 蜜臀久久99精品久久宅男| 精品免费久久久久久久清纯| 亚洲图色成人| 久久久久国产网址| 日韩欧美一区二区三区在线观看| 插阴视频在线观看视频| 联通29元200g的流量卡| 99国产极品粉嫩在线观看| 久久6这里有精品| 欧美高清性xxxxhd video| 一个人看的www免费观看视频| 国产精品久久久久久精品电影| 午夜激情福利司机影院| av.在线天堂| 精品免费久久久久久久清纯| 久久久欧美国产精品| 亚洲熟妇中文字幕五十中出| 国内久久婷婷六月综合欲色啪| 亚洲精品日韩av片在线观看| 免费在线观看成人毛片| 一进一出抽搐gif免费好疼| 能在线免费看毛片的网站| 九草在线视频观看| 超碰av人人做人人爽久久| 免费一级毛片在线播放高清视频| 免费不卡的大黄色大毛片视频在线观看 | 乱人视频在线观看| 亚洲丝袜综合中文字幕| 亚洲av男天堂| 蜜桃久久精品国产亚洲av| 深夜a级毛片| 99热6这里只有精品| 亚洲一区二区三区色噜噜| 一级毛片我不卡| 人妻制服诱惑在线中文字幕| 18禁在线播放成人免费| 午夜免费男女啪啪视频观看| 哪里可以看免费的av片| 男女那种视频在线观看| 日日撸夜夜添| 国产精品永久免费网站| 精华霜和精华液先用哪个| 国产日本99.免费观看| 国产精品综合久久久久久久免费| 国内精品宾馆在线| 美女脱内裤让男人舔精品视频 | 国产探花极品一区二区| 天堂√8在线中文| 国产精品蜜桃在线观看 | 别揉我奶头 嗯啊视频| 国产老妇女一区| 亚洲精品久久久久久婷婷小说 | 91麻豆精品激情在线观看国产| 国产人妻一区二区三区在| 免费无遮挡裸体视频| 99热只有精品国产| 狂野欧美白嫩少妇大欣赏| 能在线免费观看的黄片| 亚洲第一区二区三区不卡| 欧美另类亚洲清纯唯美| 亚洲av中文av极速乱| 哪里可以看免费的av片| 欧美色欧美亚洲另类二区| 波野结衣二区三区在线| av免费在线看不卡| 精品欧美国产一区二区三| 赤兔流量卡办理| 黄片无遮挡物在线观看| 午夜老司机福利剧场| 精品久久久久久久人妻蜜臀av| 村上凉子中文字幕在线| 99riav亚洲国产免费| 久久久欧美国产精品| 日韩精品有码人妻一区| 观看美女的网站| 成人午夜高清在线视频| av黄色大香蕉| 国产真实伦视频高清在线观看| 午夜爱爱视频在线播放| 人体艺术视频欧美日本| 免费观看人在逋| 啦啦啦观看免费观看视频高清| 精华霜和精华液先用哪个| 成人性生交大片免费视频hd| 亚洲成人av在线免费| 久久久久久久午夜电影| 欧美另类亚洲清纯唯美| 麻豆久久精品国产亚洲av| 国产av不卡久久| 成年av动漫网址| 永久网站在线| 国产黄色小视频在线观看| av在线老鸭窝| 日本色播在线视频| 久99久视频精品免费| 成人高潮视频无遮挡免费网站| 色哟哟·www| 女人十人毛片免费观看3o分钟| 久久精品久久久久久久性| 亚洲人成网站高清观看| 一区福利在线观看| 免费在线观看成人毛片| 亚洲欧美精品自产自拍| 日日摸夜夜添夜夜爱| 青春草视频在线免费观看| 国产蜜桃级精品一区二区三区| 国产精品一二三区在线看| 欧美三级亚洲精品| 午夜亚洲福利在线播放| a级一级毛片免费在线观看| 久久精品国产亚洲av香蕉五月| 干丝袜人妻中文字幕| 欧美+亚洲+日韩+国产| 国产老妇伦熟女老妇高清| 亚洲中文字幕日韩| 天天躁日日操中文字幕| 亚洲,欧美,日韩| 色综合站精品国产| 午夜久久久久精精品| 中文字幕av成人在线电影| 69人妻影院| 国产成人精品一,二区 | 日韩av在线大香蕉| 成人美女网站在线观看视频| 久久99精品国语久久久| videossex国产| 中国美白少妇内射xxxbb| 高清日韩中文字幕在线| 国产69精品久久久久777片| 成人毛片60女人毛片免费| 欧美xxxx黑人xx丫x性爽| 少妇熟女aⅴ在线视频| 免费av毛片视频| 国产一区二区在线av高清观看| 超碰av人人做人人爽久久| 91狼人影院| 亚洲国产欧美人成| 99久久成人亚洲精品观看| 亚洲成人中文字幕在线播放| 国产中年淑女户外野战色| 国内精品美女久久久久久| 18禁在线无遮挡免费观看视频| 成人永久免费在线观看视频| 国产一区二区在线av高清观看| 男人和女人高潮做爰伦理| av专区在线播放| av免费在线看不卡| 国内久久婷婷六月综合欲色啪| 久久国产乱子免费精品| 午夜精品在线福利| 麻豆乱淫一区二区| 亚洲人成网站在线播放欧美日韩| а√天堂www在线а√下载| 成人午夜精彩视频在线观看| 高清日韩中文字幕在线| 国产一级毛片在线| 日韩av不卡免费在线播放| 九九热线精品视视频播放| 看免费成人av毛片| 精品人妻视频免费看| 给我免费播放毛片高清在线观看| 欧美日韩综合久久久久久| 69av精品久久久久久| 亚洲精华国产精华液的使用体验 | 亚洲四区av| 夜夜夜夜夜久久久久| 日韩视频在线欧美| 国产午夜精品久久久久久一区二区三区| 1024手机看黄色片| www.av在线官网国产| 青春草亚洲视频在线观看| 成人一区二区视频在线观看| 免费电影在线观看免费观看| 欧美另类亚洲清纯唯美| 国产 一区精品| 日韩一区二区三区影片| 亚洲成人久久性| 18禁裸乳无遮挡免费网站照片| 99久久九九国产精品国产免费| 在线观看美女被高潮喷水网站| 日本成人三级电影网站| 国产高清视频在线观看网站| 国产亚洲精品av在线| 丝袜美腿在线中文| 成年版毛片免费区| 色播亚洲综合网| 伦理电影大哥的女人| 男女边吃奶边做爰视频| 精品免费久久久久久久清纯| 欧美潮喷喷水| 久久精品国产清高在天天线| 成年版毛片免费区| 99九九线精品视频在线观看视频| 久久久欧美国产精品| av又黄又爽大尺度在线免费看 | 免费观看人在逋| 欧美激情久久久久久爽电影| 激情 狠狠 欧美| 联通29元200g的流量卡| 在线免费观看的www视频| 久久久久久久久久久丰满| 久久99热这里只有精品18| 国产一级毛片在线| 欧美日韩在线观看h| 熟女人妻精品中文字幕| 亚洲第一区二区三区不卡| 久久久久九九精品影院| 国产黄片美女视频| 最近中文字幕高清免费大全6| 中国美白少妇内射xxxbb| 一级毛片电影观看 | 中文字幕免费在线视频6| 国产极品天堂在线| 国内精品宾馆在线| 人人妻人人澡人人爽人人夜夜 | 国产成人a区在线观看| 日韩,欧美,国产一区二区三区 | 亚洲精品国产成人久久av| 两个人视频免费观看高清| 黄色配什么色好看| 亚洲欧美日韩卡通动漫| 99久久九九国产精品国产免费| 青春草亚洲视频在线观看| 欧美人与善性xxx| 偷拍熟女少妇极品色| 国产伦理片在线播放av一区 | 国产中年淑女户外野战色| 国产精品福利在线免费观看| 欧美成人免费av一区二区三区| 亚洲欧美精品专区久久| 如何舔出高潮| 色婷婷av一区二区三区视频| 一级爰片在线观看| 日本vs欧美在线观看视频| 免费黄频网站在线观看国产| 日韩三级伦理在线观看| 精品一品国产午夜福利视频| 久久ye,这里只有精品| 日本黄色日本黄色录像| 毛片一级片免费看久久久久| 亚洲经典国产精华液单| 能在线免费看毛片的网站| 欧美三级亚洲精品| a 毛片基地| a级毛片在线看网站| 最近的中文字幕免费完整| 天天躁夜夜躁狠狠久久av| av女优亚洲男人天堂| 国产av码专区亚洲av| av电影中文网址| 亚洲av二区三区四区| 久久韩国三级中文字幕| 久久精品国产亚洲av涩爱| 国产成人av激情在线播放 | 七月丁香在线播放| 日日摸夜夜添夜夜添av毛片| av电影中文网址| 汤姆久久久久久久影院中文字幕| 十八禁高潮呻吟视频| 伦理电影免费视频| 99国产综合亚洲精品| 亚洲精品第二区| 亚洲美女视频黄频| 久久久欧美国产精品| videossex国产| 最新的欧美精品一区二区| 老熟女久久久| 制服人妻中文乱码| 在线天堂最新版资源| 久久99热6这里只有精品| 成人国语在线视频| 人人妻人人爽人人添夜夜欢视频| 我要看黄色一级片免费的| 汤姆久久久久久久影院中文字幕| 国产亚洲av片在线观看秒播厂| 亚洲国产精品一区三区| 肉色欧美久久久久久久蜜桃| 999精品在线视频| 中文精品一卡2卡3卡4更新| 久久久久国产网址| 国产精品99久久99久久久不卡 | 中文乱码字字幕精品一区二区三区| 免费看av在线观看网站| 亚洲av欧美aⅴ国产| 国产精品一区www在线观看| 日韩一本色道免费dvd| 热re99久久精品国产66热6| 制服诱惑二区| 国产欧美日韩一区二区三区在线 | 亚洲av免费高清在线观看| 永久网站在线| 51国产日韩欧美| 在线看a的网站| 男人爽女人下面视频在线观看| 久久99一区二区三区| 免费观看的影片在线观看| 午夜福利网站1000一区二区三区| 精品人妻熟女av久视频| 久久久午夜欧美精品| 一边摸一边做爽爽视频免费| 亚洲精品日韩在线中文字幕| 啦啦啦啦在线视频资源| 国产精品秋霞免费鲁丝片| 欧美激情 高清一区二区三区| 免费观看的影片在线观看| 69精品国产乱码久久久| 国产av一区二区精品久久| 日韩欧美精品免费久久| 如日韩欧美国产精品一区二区三区 | 久久婷婷青草| 亚洲婷婷狠狠爱综合网| 国产午夜精品久久久久久一区二区三区| 精品一区二区三卡| 国产精品久久久久久av不卡| 欧美日韩成人在线一区二区| 午夜影院在线不卡| 久久韩国三级中文字幕| 久久99精品国语久久久| 男女无遮挡免费网站观看| 亚洲高清免费不卡视频| 欧美日韩在线观看h| av专区在线播放| 日韩欧美精品免费久久| 观看av在线不卡| 久久久欧美国产精品| 欧美 亚洲 国产 日韩一| 日韩伦理黄色片| 日韩成人伦理影院| 老司机影院成人| 亚洲怡红院男人天堂| 亚洲在久久综合| 人妻少妇偷人精品九色| 99久久中文字幕三级久久日本| 免费大片黄手机在线观看| 韩国av在线不卡| 啦啦啦在线观看免费高清www| 久久精品人人爽人人爽视色| 韩国高清视频一区二区三区| 精品久久久久久久久av| 国产有黄有色有爽视频| 特大巨黑吊av在线直播|