• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Measuring loblolly pine crowns with drone imagery through deep learning

    2022-02-26 10:15:16XiongweiLouYanxiaoHuangLumingFangSiqiHuangHailiGaoLaibangYangYuhuiWenguaiHung
    Journal of Forestry Research 2022年1期

    Xiongwei Lou · Yanxiao Huang · Luming Fang · Siqi Huang · Haili Gao · Laibang Yang · Yuhui Weng · I.-K.uai Hung

    Abstract In modeling forest stand growth and yield, crown width, a measure for stand density, is among the parameters that allows for estimating stand timber volumes.However, accurately measuring tree crown size in the field, in particular for mature trees, is challenging.This study demonstrated a novel method of applying machine learning algorithms to aerial imagery acquired by an unmanned aerial vehicle (UAV) to identify tree crowns and their widths in two loblolly pine plantations in eastern Texas, USA.An ortho mosaic image derived from UAV-captured aerial photos was acquired for each plantation (a young stand before canopy closure, a mature stand with a closed canopy).For each site, the images were split into two subsets: one for training and one for validation purposes.Three widely used object detection methods in deep learning, the Faster region-based convolutional neural network (Faster R-CNN), You Only Look Once version 3 (YOLOv3), and single shot detection (SSD), were applied to the training data, respectively.Each was used to train the model for performing crown recognition and crown extraction.Each model output was evaluated using an independent test data set.All three models were successful in detecting tree crowns with an accuracy greater than 93%, except the Faster R-CNN model that failed on the mature site.On the young site, the SSD model performed the best for crown extraction with a coefficient of determination (R2) of 0.92, followed by Faster R-CNN (0.88) and YOLOv3 (0.62).As to the mature site, the SSD model achieved a R2 as high as 0.94, follow by YOLOv3 (0.69).These deep leaning algorithms, in particular the SSD model, proved to be successfully in identifying tree crowns and estimating crown widths with satisfactory accuracy.For the purpose of forest inventory on loblolly pine plantations, using UAV-captured imagery paired with the SSD object detention application is a cost-effective alternative to traditional ground measurement.

    Keywords UAV image · Crown recognition · Object detection · Crown width measurement

    Introduction

    The tree crown consists of a mass of foliage and branches, the component important for photosynthesis.Therefore, tree crown size (crown width and depth) is strongly correlated with the growth of a tree (Chubey et al.2006).Tree crown measurement is used to assist in estimating biomass, predicting growth, and determining wood properties (Amarasekara and Denne 2002; Popescu et al.2003; Dobbertin 2005).Although tree crown depth, a part of the tree height, can be measured using a hypsometer with acceptable accuracy, measuring the crown width is often a challenge.Conventionally, crown width is measured in the field using logger’s tape by measuring a tree’s crown diameter twice (the largest and smallest or the east-west and north-south diameters) and calculating their average as the tree crown width (Lee et al.2010).The method is widely used, although it is time-consuming and labor intense.However, its accuracy is often compromised due to terrain and ground cover.As a result, the efficacy and accuracy of techniques to measure tree crown width continue to be improved.

    Modern remote sensing technologies to identify and measure tree crowns (Ke and Quackenbush 2011) have included aerial imagery, high-resolution satellite imagery, and light detection and ranging (LiDAR).Using the Taguchi method, Erfanifard et al.(2014) optimized the support vector machines (SVM) classification system to identify tree crowns.Based on the UltraCam-D (UCD) aerial images with a spatial resolution of 0.06 m, they reported an overall accuracy of crown classification of 97.7%.Wagner et al.(2018) reported a method of automatic tree crown delineation using very high-resolution satellite images, while Li et al.(2019) proposed a real-time tree crown detection method for largescale remote sensing images.While satellite remote sensing images typically cover a larger area compared to aerial imagery, they are affected by atmospheric conditions and cloud cover.In addition, they might not provide enough spatial resolution for individual tree crown classification.Harikumar et al.(2017) applied high-density LiDAR point cloud data to build the internal and external geometric structures of a coniferous tree crown and identify tree species.Naveed et al.(2019) proposed a framework to integrate spectral and height information for individual tree crowns (ITC) delineation, with the treetop segmented to recognize tree crown.However, these methods using LiDAR and multispectral remote sensing data can be costly.

    Most of the methods developed to improve accuracy of crown width measurement without taking measurements on the ground are geometry-based, e.g., determination of the crown vertex, multiscale segmentation, or direct detection of the crown contour.Few methods have involved object recognition using the image directly.By using optical submeter resolution satellite images, Gomes et al.(2018) developed a new algorithm for the detection and delineation of tree crowns.The algorithm uses a three-dimensional geometric model of the crown to detect and calculate the crown width.Wu et al.(2016) captured the topographical structure of forest and quantified the topological relationships of tree crowns using a graph theory-based localized contour tree method, before segmenting individual tree crowns analogous to recognizing hills from a topographic map, thus improving the accuracy by 1.8%-4.5% over that for the watershed and local maximum algorithm.Based on simple linear iterative clustering of superpixel algorithm, Zhou et al.(2020) obtained the crown boundaries based on aerial images of forest plots captured by drones.However, the methods listed above can only be applied to measure tree crowns at specific terrains and their accuracies are expected to be lower when applied to forest stands of multiple canopy layers.

    As reviewed above, image-based crown width measurement involves crown classification (detection) and measurement.In recent years, deep learning based object detection algorithms, which can classify and measure objects in images have been widely used in fields such as in medical science (Kumar et al.2016; Wu et al.2018) and facial recognition (He et al.2018).Ren et al.(2018) was able to detect aircraft and ships by improving the Faster R-CNN algorithm and using high-resolution optical remote sensing images.Deep learning has also been applied to forestry.For example, Santos et al.(2019) employed drones to take images over a forested urban area in midwest Brazil, and applied three different object detection algorithms directly to recognize tree species.With the rapid development in deep learning, a subset of machine learning in artificial intelligence, the use of UAV with object detection algorithms may be an efficient alternative to measuring tree crowns, but has not been demonstrated yet.

    In this study, deep learning based objection detection algorithms were exercised to identify and measure individual tree crown width of pine plantations based on highresolution images captured by UAV, with the goal being to demonstrate the efficacy and accuracy of the method.Three popular algorithms, Faster R-CNN (Ren et al.2015), You Only Look Once version 3 (YOLOv3) (Redmon and Farhadi 2018), and Single Shot Detection (SSD) (Liu et al.2016), were applied.Given the increasing popularity of drone and machine learning techniques in forestry, results of the study may have significant impact in future forest mensuration.

    Object detection using deep learning

    For the development of a typical statistical model, organized data are run through predefined equations, then model parameters are estimated, yielding predictive system outputs.Alternatively, a deep learning process trains the computer to learn on its own by recognizing patterns in the data using multiple layers of processing.Most recent deep learning models are based on artificial neural networks, specifically, convolutional neural networks (CNNs) to recognize latent features of the data from regularities in the layers.If the output data resulted from this network (model) process complies with the model requirements (LeCun et al.2015), the network will be retained as the target model.Otherwise, the parameter setting of the network will be readjusted iteratively until the output meets the requirements.

    Deep learning has been extensively and successfully applied in many areas, including object detection.Object detection provides a fast and accurate means to classify and position (measure) objects in images (Girshick et al.2014).An object detection model predicts bounding boxes with confidence scores, one for each object it finds, as well as classification probabilities for each object.Object classification judges whether the object of interest is in the input image and reports a series of scored labels, which indicate the probability of the object of interest being in the input image.The object positioning determines the position and range of the object of interest in the input image.Then it delineates the border or the center of the object, a square box known as a bounding box, which can be used to measure object size.For this study, we used three commonly used algorithms: Faster R-CNN (Ren et al.2015), You Only Look Once version 3 (YOLOv3) (Redmon and Farhadi 2018), and single shot detection (SSD) (Liu et al.2016).

    Faster R?CNN

    Faster R-CNN is a two-stage detection algorithm, with the first stage identifying a subset of regions in an image that might contain an object and the second stage classifying the object in each region.This algorithm consists of two modules: a deep fully convolutional network that proposes regions, and a Faster R-CNN detector that processes the proposed regions.It utilizes a region proposal network (RPN) with the CNN model.The RPN shares full-image convolutional features with the detection network and can simultaneously estimate object bounds and object scores at each position.Application of the technique takes following steps: pass an input image to the ConvNet, which returns feature maps for the image; apply a RPN on these feature maps and receive object proposals; apply ROI pooling layer to bring down all the proposals to the same size (Girshick 2015); and pass these proposals to a fully connected layer to classify any bounding boxes for the image.The Faster R-CNN detectors are designed to improve detection performance with a large number of regions.

    YOLOv3

    YOLOv3, an object detection system targeted for real-time processing and based on the structure of Feature Pyramid Networks (FPN) (Lin et al.2014), uses a single stage object detection network that is faster than other two-stage deep learning object detectors, such as Faster R-CNNs.The YOLOv3 model runs a deep learning CNN on an input image to produce network predictions.The object detector decodes the predictions and generates bounding boxes.YOLOv3 uses a feature extractor that has 53 convolutional layers and 22 Res layers named Darknet53 that is trained on ImageNet.Res layers use the idea of Residual Network (ResNet) (He et al.2016) for reference.Seventy-five layers are stacked to the feature extractor resulting feature fusion of three different levels of feature maps, so that the detector can better detect objects of different sizes.Development of a YOLOv3 model involves (1) inputting the image data for training, (2) annotating label objectives on images and initiating training, and (3) testing the trained model with YOLOv3.

    SSD

    SSD is another single-stage object detection model, and it adopts different detection steps from YOLOv3.After SSD extracts features from CNN to generate a feature map, it selects six layers of different sizes (multi-scale) from feature maps to perform straightforward regression for generating not only a series of fixed size bounding boxes, but also the confidence level for each object in these bounding boxes, before non-maximum suppression (NMS) (Neubeck and Van Gool 2006) is applied to produce the final detection results.To reduce the number of parameters and the computing workload, SSD adopts the MobilenetV2 (Sandler et al.2018) network as the basic network.The MobilenetV2 network inherits the highly efficient deep separable convolution form and the Residual Network in MobilenetV1 (Howard et al.2017), based on which a linear ResNet is proposed to increase the dimensionality of the input data.Application of SSD includes the following steps: install labelling, using labelling to annotate photos, config SSD, train model, and test model.

    Materials and methods

    Study sites and image acquisition

    Two loblolly pine (Pinus taedaL.) plantations were purposely selected with one representing a young stand and the other a mature stand.The young stand was established in Houston County, Texas (31°18′38.0? N, 95°18′07.1? W) in 2015 with a spacing of 2.4 m by 2.7 m (1,500 trees per hectare).In 2019, the plantation had a good survival rate (92%) and reasonable growth, averaging 3.5 m in total height and 8 cm in diameter at breast height (DBH).This site has a flat terrain, and its tree crowns were not closed at the time of the study.The mature stand is in Rusk County, Texas (31°45′31.3? N, 95°02′31.8? W).The site is also flat and was an old field before it was planted with loblolly pine in 2001.In 2019, the site averaged 22.2 cm in DBH, 16.9 m in total height, and 35.2 m2in basal area per hectare, where the canopy was closed.

    A DJI Phantom 4 Pro produced by Shenzhen DJI Sciences and Technologies Ltd.was used to capture aerial images of the two study sites, on 5 November 2019 for the young stand and 13 September 2019 for the mature stand.The Pix4Dcapyure app was used on a tablet to plan each flight mission, assigning parameters including fly area, image overlap, angle of camera, flight speed and altitude, etc.The flight altitude of the UAV was set to 42 m for the young site and 46 m for the mature site, both with the camera facing down vertically at 90°.

    Individual images for each site were aligned to generate a dense point cloud and high-resolution ortho rectified image using Agisoft Photoscan (2014), a software application which performs photogrammetric processing of digital images (i.e., aerial triangulation and bundle block adjustment) and generates 3D spatial data to be used for indirect measurements of objects of various scales.The image size of ortho mosaic for the young site was 20,886 × 19,445 pixels, with the spatial resolution of 1.1720 cm per pixel, while the mature site has an image size of 28,376 × 22,244 pixels and 1.2446 cm resolution.Figure 1 shows the two site locations and their respective ortho mosaic images.

    Measurement of actual tree crown width

    Conventionally, tree crown width should be measured in the field.We thus attempted field measurement at both sites, but the results were less than satisfactory.At the young site, grasses were too dense for us to approach the trees.At the mature stand where the canopy was closed, the tree crowns were too intertwined for exact tree crown edges to be determined, particularly when looking into direct sunlight.Instead, we measured the crown width of each tree based on the ortho mosaic image in two directions (longest width and shorted width) and calculated their average as the actual tree crown width (hereafter referred to as actual crown width).This procedure is labor intense but is believed to result in more accurate values than field measurements.

    Object detection model development

    In this study, Faster R-CNN, YOLOv3 and SSD algorithms were used to develop models for measuring tree crown width, which involved both crown classification and regression.Classification evaluates the model accuracy in identifying tree crowns, while regression evaluates the model accuracy in measuring tree crown width.

    Fig.1 The two study sites of loblolly pine plantation located in east Texas.Note: Each red square (a) indicates the location of a study site.Close-ups of the young site (b) and mature site (d) are shown on a Google satellite map and viewed on the ortho mosaic image, young (c) and mature (e)

    For the young site, 128 trees within a plot that was randomly selected pre-flight were used for the model test (test data set) (Fig.2).The remaining part of the image was randomly sampled to create the training data set.In total, 445 samples were created, with each sample cropped into an image subset of 500 × 500 pixels that contained multiple tree crowns.For the mature site, where no test data set was selected before the flight, a total of 207 samples were randomly selected from the entire ortho mosaic image as the training data, with each sample also being cropped into an image subset of 500 × 500 pixels.Another 185 trees, which were independent of the training data, were extracted as the test data set.

    Fig.2 Aerial image of the young site showing the 128 tress used for model test

    The training data set for each site was input to the three models, Faster R-CNN, YOLOv3 and SSD, respectively.During the training process, the data set was further divided into two subsets, a training subset and a validation subset, at a 9:1 ratio for iteration.These samples were annotated using LabelImg software (https:// github.com/ tzuta lin/ label Img), where each canopy within a sample was marked with a rectangular border manually (Fig.3).The parameters were set as follows for training Faster R-CNN, YOLOv3 and SSD: batch sizes were 8, 8, 16; learning rates were 0.0001, 0.001, 0.004; epochs were 720, 120 and 500, respectively.After the training of a certain number of epochs, the loss values of the models were stabilized in the training and the validation subsets.The three network models were run on the Tensorflow (Abadi et al.2016) platform of the Windows 10 operating system.The computer has an Intel(R) Xeon(R) Silver 4110 CPU with a clock rate of 2.10 GHz, 64 GB RAM, and a NVIDIA GeForce GTX 1080 Ti graphic processor with 11 GB memory.

    Fig.3 Crown areas in the random samples were manually delineated using labelIng software (https:// github.com/ tzuta lin/ label Img) for model training in the two study sites

    Faster R-CNN model requires a large amount of data to improve both robustness (Bertsimas and Sim 2004) and accuracy.However, the experimental, labeled data for both the study sites were limited.Transfer learning is helpful for mitigating this limitation (Pan and Young 2009).Specifically, training is performed on a large data set (such as COCO data set; Lin et al.2014), and then the resulting weights are taken as the initial parameters of the training model.Of course, not all weights need to be retrained, only the weights of the last layer are trained, such as the fully connected layer.This process is referred to as fine-tuning.Data augmentation provides an alternative way to avoid model overfitting, where the original sample is horizontally flipped and randomly cropped to increase the number of data sets (Perez and Wang 2017).We applied both methods to Faster R-CNN, and the results enabled Faster R-CNN to extract crown from the young site, but failed for the mature site, probably due to its dense canopy.Therefore, the detection results of Faster R-CNN for the mature site were not acceptable and thus not reported in this study.

    Model evaluation

    The test data were used to assess the accuracy of the tree crown classification and crown width detection.The classification indexes were calculated to evaluate whether the crowns were correctly identified, while the regression indexes were used to evaluate the differences between the estimated crown size and the actual crown size.

    Intersection over union (IoU) calculates the ratio of the intersection and union between the bounding box identified by the model and the real object bounding box.The object detection task does not require that the identified bounding box exactly match the real one.If an IoU is equal to or greater than 0.5, the detection is considered as correct (Everingham et al.2015).

    Four parameters are used to quantify the correctness of each model output: true positive (TP), the number of correctly identified tree crowns in an image; false positive (FP), the number of wrongly identified crowns; false negative (FN), the number of unidentified crowns; and true negative (TN), the number of correctly identified non-crown objects.The model classification performance was evaluated based on precision, recall, accuracy and F1 score (Eqs.1-5).Precision is the ratio of correctly identified crowns over all identified crowns and represents the misidentification of the crown in the detection result.Recall represents the recall effect of classification, representing the correct frequency in the case of positive samples and shows the missed recognition of the crown in the test results.Accuracy represents the ratio of the number of correctly classified samples to the total number of samples.Note that in our study, only the crown is detected as a positive class, and thus TN=0, which simplifies the calculation of the accuracy.F1 Score is the harmonic mean of precision and recall.The two indicators accuracy and F1 score provide an overall evaluation of the model.

    The model accuracy in measuring tree crown width was evaluated by calculating model bias, root mean square error (RMSE), and coefficient of determination (R2) (Eqs.6-8).Bias represents the degree of deviation between the estimated and actual values, RMSE andR2represent model precision.

    whereNrepresents the number of samples, yiindicates the actual value,denotes the estimated value, and

    Results

    Overall effects

    By applying the test data to the models built upon the training data, the models were able to delimit individual tree crown frames and identify crown size with confidence (Figs.4, 5).Through visualization, we found that crown detection was successful throughout the entire study site, both young and mature, regardless of the object detection model used, or whether it was within or outside the areas for the test data set.

    Fig.4 Results of crown identification at the young site using the three models.a Faster R-CNN, b YOLOv3, and c SSD.Note: Images d-f are the respective close-ups of area boxed in red in images a-c

    Fig.5 Crown identification results for the two models using the entire ortho mosaic image of the mature site: a YOLOv3b and b SSD.The red boxes in a and b mark all validation areas, which are shown in close-ups in c and d; the red boxes in c and d show the bounding boxes that resulted from each model

    As expected, Faster R-CNN took the longest time (68 ms) for the detection since it utilizes a two-stage network.The other two models use a one-stage network, and SSD applied MobilenetV2 further as the basic network to reduce the computation workload and improve the recognition speed at 25 ms, about half the time needed by YOLOv3 (49 ms).

    Faster R-CNN first generates region proposals, then makes a separate estimation for each of these regions.Therefore, it took more time because it requires running the detection and classification portion of the model multiple times.On the other hand, one-stage detectors such as YOLOv3 and SSD take a one-stage strategy by assigning each bounding box detector to a specific position in the image and are expected to be faster than Faster R-CNN.

    Crown detection

    Table 1 shows the results of model classification of tree crowns by applying the test data to the trained models for the young and mature sites.For all models and both sites, all of the classification indexes, including precision, recall, accuracy, and F1, achieved at least 93%, an indication that all of the three models were successful in identifying crowns in the young and the mature sites, except the Faster R-CNN model, which failed on the mature site.

    Table 1 Evaluation of crown classification for the young and mature forests by the three models

    For the young site, almost all trees were correctly identified (precision) by the Faster R-CNN (99.22%) and SSD (99.21%) models.The precision of YOLOv3 (95.45%) was also high, although slightly lower (3.75%) than the Faster R-CNN and SSD.For the mature site, the YOLOv3 (98.31%) and SSD (97.83%) were similar with high precision.The recall parameter represents the missing recognition of treecrown in the detection results.For the young site, the recall of Faster R-CNN was the highest possible (100%), indicating that every tree was identified, 1.56% and 2.34% higher than that of YOLOv3 (98.44%) and SSD (97.66%), respectively.For the mature site, however, SSD (97.30%) had a 2.71% higher recall than that of YOLOv3 (94.59%).YOLOv3 and SSD had a similar likelihood of misidentification in the young site where objects were sparse.However, for the mature site with dense canopy, SSD had slightly better recognition than YOLOv3 did.

    For the young site, the accuracy of Faster R-CNN reached 99.22%, suggesting this model identified the crowns satisfactorily.The YOLOv3 (94.03%) and SSD (96.90%) models had somewhat lower accuracy but still considered to be accurate.Nearly the same high level of accuracy for the young site was also obtained for the mature site using SSD (95.24%) and YOLOv3 (93.09%).The F1 score provides an overall evaluation by combining accuracy and precision of the model into one parameter (Goutte and Gaussier 2005).According to the F1 score, the ranking of the three models started with Faster R-CNN (99.61%), followed by SSD (98.43%) and YOLOv3 (96.92%) for the young site, and SSD (97.56%) was also better than YOLOv3 (96.42%) for the mature site.

    Figure 6 shows the model precision-recall (P-R) curves based on the validation data.A certain point on the P-R curve (Davis and Goadrich 2006) represents that under a certain threshold, the model judges any result with confidence greater than the threshold as a positive sample, and any result less than the threshold as a negative sample.At this time, the corresponding precision and recall are determined.The entire P-R curve is generated by moving the threshold from high to low.If the P-R curve of one learner is completely covered by the curve of another learner, the performance of the latter is considered superior to the former.In the young site, Faster R-CNN and SSD performed similarly, and both had a detection effect significantly better than YOLOv3.For the mature site, the two curves (YOLOv3 and SSD) intersect.At this point, the break even point (BEP) is adopted as a measure of the model evaluation.BEP is defined as the value when the P-R curve intersects a line of 100% slope where precision=recall.The BEP value of SSD was slightly higher than that of YOLOv3, implying that the effect of SSD crown identification is slightly better than that of YOLOv3.

    Fig.6 The precision-recall (P-R) curves using different models for a the young site and b mature site

    Crown width estimation

    For a correctly identified crown, its width can be estimated by calculating its bounding box dimension.By applying the test data to each model, the results (Table 2, Fig.7) showed that the model bias and precision in estimating crown width varied between models and between site types.For the young site, the SSD model had the lowest bias (0.44) and highest precision with RMSE=0.17 andR2=0.92, which translated to higher accuracy.Although its superiority to Faster R-CNN in both terms was not substantial, it was significantly better than YOLOv3.For the mature site, theR2value for SSD reached 0.94, which was much higher than for YOLOv3 (0.69), although both models displayed similar bias values.TheR2values for SSD and YOLOv3 were slightly higher for the mature site than for the young forest.Overall, the results suggest that the SSD model is more effective than the YOLOv3 model in extracting the true boundary of the tree crown.The Faster R-CNN model was also a good alternative for the young site, where the tree canopy was not closed.Figure 7 also displays how each model under-or over-estimated tree crown width.All models estimated tree crown width well at the lower end, but underestimated width at the higher end of tree crown size.Comparatively, the SSD model estimated tree crown width at both ends better than the other two models.

    Fig.7 Comparison between the estimated and actual crown widths determined by the different models for the a-c young site and d, e mature site

    Table 2 Assessment of crown width estimation by different models

    Discussion

    Faster R-CNN is a two-stage network and was expected to perform better than the one-stage algorithms, SSD and YOLOv3.The results confirmed this expectation for the young site with the open canopy.However, it failed toproduce acceptable results for the mature site perhaps as a result of the closed canopy at the mature site, where more samples might be needed for the Faster R-CNN algorithm to be successful.More studies are required to further investigate this issue.SSD performed better than YOLOv3 on both sites, particularly for crown width estimation.With Darknet-53 as the backbone, YOLOv3 had fast training speed and high versatility.However, the number of iterations is fxied in pursuit of higher speed, which leads to lower accuracy than SSD for detecting objects.Based on the default proposal box mechanism of the Faster R-CNN network and the regression mechanism of YOLO (Redmon et al.2016), SSD adopts the idea of multi-scale feature characterization to detect the object.Thus, it combines the fast speed of YOLO and the high accuracy of Faster R-CNN, a breakthrough for the existing detection framework.

    Conventionally, the width of a tree crown is measured by projecting the edges of the crown to the ground (dripline) and measuring the length along one axis from edge to edge through the crown center.Since a crown usually has an irregular shape, two axes (the max or min axis, or SW and NE axis) are normally measured and averaged as the crown width.Accurate projecting the crown edges to the ground is difficult, in particular for mature plantations with closed canopy.Thus, the ground measurement may be accurate for young and short trees, but tends to be biased for mature trees in rough terrains and complex stands.In addition, this method is labor intense and requires at least two crew members.Compared to the conventional method, the object detection models described in this study greatly mitigate the field workload; most work can be completed with a computer.Based on the results, even in the mature stand with a closed canopy, the models correctly bound and measured tree crown width (Figs.5, 7).Our models were also superior to other methods such as conventional aerial photogrammetry or remote sensing, which are generally suited for detecting large-area targets.With respect to forestry, they are commonly used to identify tree species in a massive area.However, the accuracy of crown extraction and the recognition of small targets could fall short of expectation.Since UAVs can obtain imagery with much higher spatial resolution than conventional aerial and satellite imagery, it is more suitable for identifying smaller targets such as individual trees.However, due to limitations in flight speed and duration, it is not suitable for large area target recognition.

    The application of this study involves three steps: flying a UAV to capture images, processing images to attain an ortho mosaic image, and applying an object detection algorism to classify tree crowns and estimate each crown width.The technology for the first two steps are in place, and their application is straight forward.The third step only assigns the image path of the ortho mosaic to the model.In return, the model generates the output with images marked with crown boundaries as shown in Fig.4 and a list of items corresponding to the identified crowns in a text document.The list includes the number, center position, left and right width, top and bottom width, and average width of each crown identified.Since the models based on deep learning techniques are data oriented, the models developed in this study may apply to other loblolly pine plantations with the similar environments.Both study sites in this study are even-aged pine plantations, where trees are evenly positioned.The efficacy of these models when applied to mixed or uneven-aged stands needs to be confirmed.Through the above analysis, we know that the detection outcome of the SSD model was the most desirable.

    The SSD model built in this study can be shared with other researchers for further assessment.If there are enough samples, the Faster R-CNN model can be adopted to improve precision through training.In addition, research can be conducted on the embedded devices that can extract tree crowns in real time.The Faster R-CNN has high accuracy although its recognition speed is low, which is ideal for desktop computers with greater computing power.In comparison, the MobilenetV2-based SSD and YOLOv3 can be used with a mobile terminal.

    Conclusions

    This work demonstrated the potential application of UAVacquired imagery paired with object detection algorithms, Faster R-CNN, YOLOv3, and SSD to estimate loblolly pine tree crowns.The results were promising, in particular for the SSD model, which achieved high accuracy for tree crown detection and crown width estimation for the young and the mature sites.In addition to accuracy, these methods also save on time and cost.A data enhancement method was used in model training to increase the number of data sets by flipping the original samples horizontally and randomly, reducing the possibility of model over-fitting, and improved model generalization and accuracy.Finally, our study contributed to the improvement of the network model.In the SSD model, the basic network was changed from VGG16 to MobilenetV2.Experiments show that this algorithm is more suitable for tree crown recognition, and the detection speed and accuracy of the model are good.Nonetheless, this is one of the first studies using deep learning technology in forestry, so there is great potential for further studies.We highly recommend the use of this object detection algorithm for forest inventory.As this machine learning technology is developed for more forest applications, the technology will improve and mature.

    久久午夜亚洲精品久久| 亚洲第一青青草原| 成人国产一区最新在线观看| 亚洲国产精品sss在线观看| avwww免费| 国产成人精品在线电影| 日韩一卡2卡3卡4卡2021年| 亚洲色图综合在线观看| 天天躁夜夜躁狠狠躁躁| 91精品三级在线观看| 一个人观看的视频www高清免费观看 | 免费久久久久久久精品成人欧美视频| 欧美精品啪啪一区二区三区| 久久狼人影院| 国产高清有码在线观看视频 | 黑丝袜美女国产一区| 亚洲一卡2卡3卡4卡5卡精品中文| 99精品在免费线老司机午夜| 91字幕亚洲| 亚洲精品美女久久久久99蜜臀| www.www免费av| 亚洲国产中文字幕在线视频| 一边摸一边抽搐一进一出视频| 无遮挡黄片免费观看| 91成年电影在线观看| 亚洲人成电影免费在线| 亚洲精品av麻豆狂野| 精品一区二区三区视频在线观看免费| 亚洲精品美女久久久久99蜜臀| 亚洲人成电影观看| 高潮久久久久久久久久久不卡| 久久香蕉精品热| 丰满的人妻完整版| 日韩高清综合在线| 50天的宝宝边吃奶边哭怎么回事| 一区二区三区高清视频在线| 国产精品爽爽va在线观看网站 | 成年女人毛片免费观看观看9| 9191精品国产免费久久| 日韩成人在线观看一区二区三区| 免费搜索国产男女视频| 中文字幕最新亚洲高清| 不卡一级毛片| 啦啦啦 在线观看视频| 给我免费播放毛片高清在线观看| 一级毛片精品| 此物有八面人人有两片| 91av网站免费观看| 成人三级做爰电影| 一边摸一边做爽爽视频免费| 亚洲成人国产一区在线观看| 欧美人与性动交α欧美精品济南到| 成人18禁高潮啪啪吃奶动态图| 一级黄色大片毛片| 十八禁网站免费在线| 亚洲成av人片免费观看| 国产1区2区3区精品| 18禁观看日本| 国产亚洲精品久久久久5区| 欧美黄色淫秽网站| 欧美不卡视频在线免费观看 | 午夜福利18| 久久人妻福利社区极品人妻图片| 一卡2卡三卡四卡精品乱码亚洲| 在线十欧美十亚洲十日本专区| 搡老岳熟女国产| 欧美日韩亚洲综合一区二区三区_| 99久久精品国产亚洲精品| 一区二区三区国产精品乱码| 97人妻精品一区二区三区麻豆 | 亚洲人成电影观看| 日日摸夜夜添夜夜添小说| 99久久久亚洲精品蜜臀av| 老汉色av国产亚洲站长工具| 午夜老司机福利片| 99国产精品一区二区三区| 天天躁狠狠躁夜夜躁狠狠躁| 亚洲aⅴ乱码一区二区在线播放 | 熟女少妇亚洲综合色aaa.| 成人国产综合亚洲| 色综合亚洲欧美另类图片| 免费看十八禁软件| 中文字幕av电影在线播放| 亚洲专区中文字幕在线| 香蕉丝袜av| 欧美中文日本在线观看视频| 国语自产精品视频在线第100页| 精品人妻在线不人妻| 1024视频免费在线观看| 午夜免费鲁丝| 最好的美女福利视频网| 久久久久久久久久久久大奶| 国产av在哪里看| 视频区欧美日本亚洲| 美女 人体艺术 gogo| 久久精品国产综合久久久| 色在线成人网| 99国产精品一区二区蜜桃av| 女人精品久久久久毛片| 不卡av一区二区三区| 99re在线观看精品视频| 一区二区三区精品91| 国内精品久久久久精免费| 午夜精品国产一区二区电影| 精品免费久久久久久久清纯| 日韩欧美国产在线观看| 久久婷婷成人综合色麻豆| 国产亚洲欧美在线一区二区| 国产区一区二久久| 少妇 在线观看| 日韩欧美三级三区| x7x7x7水蜜桃| 精品福利观看| 1024香蕉在线观看| 中文字幕人妻熟女乱码| 亚洲自偷自拍图片 自拍| 午夜a级毛片| www.www免费av| 欧美一级a爱片免费观看看 | 纯流量卡能插随身wifi吗| 久久精品人人爽人人爽视色| 91麻豆av在线| 欧美亚洲日本最大视频资源| 91av网站免费观看| 9色porny在线观看| 女人高潮潮喷娇喘18禁视频| 久久天躁狠狠躁夜夜2o2o| 最新美女视频免费是黄的| 一区二区三区高清视频在线| 狠狠狠狠99中文字幕| 国内久久婷婷六月综合欲色啪| 亚洲国产高清在线一区二区三 | 黄频高清免费视频| 日本三级黄在线观看| 久久久久久免费高清国产稀缺| 欧美日本中文国产一区发布| 大码成人一级视频| 国产99白浆流出| 女生性感内裤真人,穿戴方法视频| 精品一区二区三区四区五区乱码| 欧美日韩精品网址| 国产午夜精品久久久久久| 日本vs欧美在线观看视频| 757午夜福利合集在线观看| 黄色女人牲交| 国产成人一区二区三区免费视频网站| 老熟妇乱子伦视频在线观看| 变态另类丝袜制服| 亚洲成国产人片在线观看| 一二三四在线观看免费中文在| 日本精品一区二区三区蜜桃| 成年女人毛片免费观看观看9| 久久国产精品影院| 亚洲av日韩精品久久久久久密| 久久精品国产99精品国产亚洲性色 | 美女午夜性视频免费| 色精品久久人妻99蜜桃| 看片在线看免费视频| 波多野结衣av一区二区av| 免费搜索国产男女视频| 看黄色毛片网站| 久久国产精品男人的天堂亚洲| 免费看十八禁软件| av电影中文网址| 婷婷六月久久综合丁香| 可以在线观看的亚洲视频| 99国产精品一区二区三区| 亚洲激情在线av| 在线观看免费午夜福利视频| 国产精品免费视频内射| 黑丝袜美女国产一区| 亚洲国产中文字幕在线视频| 免费在线观看影片大全网站| 香蕉丝袜av| 91大片在线观看| 韩国精品一区二区三区| 久久久久国内视频| 欧美日本视频| 国产精品二区激情视频| 色av中文字幕| 亚洲全国av大片| 一夜夜www| 久久影院123| 波多野结衣高清无吗| 国产成人精品无人区| 91老司机精品| 日韩 欧美 亚洲 中文字幕| 亚洲五月婷婷丁香| 美女国产高潮福利片在线看| 久久久久久大精品| 美女扒开内裤让男人捅视频| 人人妻,人人澡人人爽秒播| cao死你这个sao货| 久久精品国产综合久久久| 男女做爰动态图高潮gif福利片 | 18禁观看日本| 午夜免费激情av| 亚洲av成人一区二区三| av视频免费观看在线观看| 国产av又大| 熟妇人妻久久中文字幕3abv| 男女床上黄色一级片免费看| 精品久久久久久成人av| 国产野战对白在线观看| 桃红色精品国产亚洲av| 欧美激情久久久久久爽电影 | 无限看片的www在线观看| 欧美老熟妇乱子伦牲交| 性欧美人与动物交配| 亚洲五月色婷婷综合| 欧美日韩中文字幕国产精品一区二区三区 | 久久精品亚洲精品国产色婷小说| 长腿黑丝高跟| 精品欧美一区二区三区在线| 欧美+亚洲+日韩+国产| 免费搜索国产男女视频| 欧美乱码精品一区二区三区| 国产成年人精品一区二区| 国产亚洲欧美98| 99国产精品一区二区三区| 国产野战对白在线观看| 精品熟女少妇八av免费久了| 国产精品乱码一区二三区的特点 | 精品国产乱子伦一区二区三区| 久久国产精品影院| 免费观看人在逋| 欧美黄色淫秽网站| 亚洲av五月六月丁香网| 黄色视频,在线免费观看| 午夜久久久久精精品| 午夜免费成人在线视频| 香蕉国产在线看| 熟妇人妻久久中文字幕3abv| 视频区欧美日本亚洲| 国产精品永久免费网站| 色在线成人网| 午夜老司机福利片| 麻豆久久精品国产亚洲av| 亚洲欧美精品综合一区二区三区| 久久中文字幕人妻熟女| 国产1区2区3区精品| 日韩欧美一区二区三区在线观看| 亚洲第一欧美日韩一区二区三区| 日本免费一区二区三区高清不卡 | 黄色成人免费大全| 免费看美女性在线毛片视频| 日日夜夜操网爽| 国产高清激情床上av| 国产成人一区二区三区免费视频网站| 999精品在线视频| cao死你这个sao货| 老司机深夜福利视频在线观看| 99国产精品免费福利视频| 老司机午夜福利在线观看视频| 身体一侧抽搐| 午夜免费激情av| 天天一区二区日本电影三级 | 成人永久免费在线观看视频| 99在线人妻在线中文字幕| 午夜成年电影在线免费观看| 国产在线观看jvid| 97超级碰碰碰精品色视频在线观看| 久久香蕉精品热| 日本a在线网址| 欧美最黄视频在线播放免费| 美女高潮喷水抽搐中文字幕| 在线播放国产精品三级| 国产一区在线观看成人免费| 97超级碰碰碰精品色视频在线观看| 最近最新中文字幕大全电影3 | 国产99白浆流出| 欧美日韩瑟瑟在线播放| 淫妇啪啪啪对白视频| 中文字幕色久视频| 99re在线观看精品视频| 波多野结衣高清无吗| 国产男靠女视频免费网站| 国产成人欧美| 正在播放国产对白刺激| 色在线成人网| 国产野战对白在线观看| 身体一侧抽搐| 男女午夜视频在线观看| 午夜激情av网站| 怎么达到女性高潮| 欧洲精品卡2卡3卡4卡5卡区| 亚洲国产精品久久男人天堂| 九色国产91popny在线| 国产精品国产高清国产av| 亚洲欧美日韩无卡精品| 成人亚洲精品一区在线观看| 日本 欧美在线| bbb黄色大片| 成人18禁在线播放| 国产成人精品久久二区二区免费| 久久香蕉精品热| 成人国产一区最新在线观看| 国产91精品成人一区二区三区| 精品国产乱码久久久久久男人| 久久久久久久久久久久大奶| aaaaa片日本免费| 欧美大码av| 亚洲国产精品合色在线| 纯流量卡能插随身wifi吗| av视频免费观看在线观看| 欧美日韩中文字幕国产精品一区二区三区 | 99精品在免费线老司机午夜| 两性午夜刺激爽爽歪歪视频在线观看 | 欧美一区二区精品小视频在线| 一区二区日韩欧美中文字幕| 亚洲中文av在线| 日韩高清综合在线| 91麻豆av在线| 9191精品国产免费久久| 午夜两性在线视频| 国产主播在线观看一区二区| 免费观看人在逋| 女人精品久久久久毛片| 夜夜躁狠狠躁天天躁| 女同久久另类99精品国产91| 欧美日本亚洲视频在线播放| 欧美黄色片欧美黄色片| 满18在线观看网站| 精品国产一区二区久久| 人妻丰满熟妇av一区二区三区| 亚洲欧美精品综合一区二区三区| 日韩成人在线观看一区二区三区| 日韩三级视频一区二区三区| 免费人成视频x8x8入口观看| 两人在一起打扑克的视频| 国产成人精品久久二区二区免费| 亚洲熟女毛片儿| av天堂在线播放| 亚洲成av片中文字幕在线观看| 国产精品av久久久久免费| 久久天躁狠狠躁夜夜2o2o| a在线观看视频网站| 很黄的视频免费| 日韩欧美国产在线观看| 中文字幕高清在线视频| av电影中文网址| 少妇粗大呻吟视频| 日韩成人在线观看一区二区三区| 色综合亚洲欧美另类图片| 51午夜福利影视在线观看| av福利片在线| 黄色丝袜av网址大全| 熟妇人妻久久中文字幕3abv| 亚洲熟女毛片儿| 精品久久久久久,| 99精品欧美一区二区三区四区| 免费女性裸体啪啪无遮挡网站| 丝袜人妻中文字幕| bbb黄色大片| 国产高清videossex| 精品久久久久久久人妻蜜臀av | 久久香蕉精品热| 色在线成人网| 欧美激情极品国产一区二区三区| 男女下面插进去视频免费观看| 禁无遮挡网站| 欧美一级毛片孕妇| 国内久久婷婷六月综合欲色啪| 啦啦啦韩国在线观看视频| 侵犯人妻中文字幕一二三四区| 啦啦啦免费观看视频1| 国产精品日韩av在线免费观看 | 国产精品电影一区二区三区| 一本综合久久免费| 亚洲精品一卡2卡三卡4卡5卡| 国产男靠女视频免费网站| 天堂√8在线中文| 香蕉丝袜av| 国产色视频综合| 最新在线观看一区二区三区| 50天的宝宝边吃奶边哭怎么回事| 欧美不卡视频在线免费观看 | 色综合欧美亚洲国产小说| 最近最新中文字幕大全电影3 | 一边摸一边抽搐一进一小说| 日本撒尿小便嘘嘘汇集6| 一级毛片精品| 久久亚洲真实| 国产亚洲欧美98| 国产伦一二天堂av在线观看| 18禁观看日本| 91麻豆精品激情在线观看国产| cao死你这个sao货| 日本撒尿小便嘘嘘汇集6| 啦啦啦韩国在线观看视频| 欧美精品啪啪一区二区三区| 丝袜在线中文字幕| 精品国产美女av久久久久小说| 一个人免费在线观看的高清视频| 一进一出抽搐动态| www日本在线高清视频| 国产91精品成人一区二区三区| 97超级碰碰碰精品色视频在线观看| 国产精品亚洲美女久久久| 国产精品久久久久久亚洲av鲁大| 午夜福利免费观看在线| 国产亚洲欧美98| 九色亚洲精品在线播放| 一二三四社区在线视频社区8| 97超级碰碰碰精品色视频在线观看| 国产av又大| 国内精品久久久久精免费| 国产av在哪里看| 久久久久久久久久久久大奶| 国产精品电影一区二区三区| 午夜福利欧美成人| 一本久久中文字幕| 国产99久久九九免费精品| 韩国精品一区二区三区| 久久中文字幕人妻熟女| 精品久久久久久成人av| 黄色 视频免费看| 叶爱在线成人免费视频播放| 制服人妻中文乱码| 男人舔女人的私密视频| 两个人视频免费观看高清| 淫妇啪啪啪对白视频| 97人妻精品一区二区三区麻豆 | 亚洲片人在线观看| 男人操女人黄网站| 精品久久久精品久久久| 久久婷婷人人爽人人干人人爱 | 久久精品人人爽人人爽视色| 久久久久久国产a免费观看| x7x7x7水蜜桃| 97人妻天天添夜夜摸| 黄色视频不卡| 性少妇av在线| 国语自产精品视频在线第100页| 黄色视频不卡| 国产精品 欧美亚洲| 国产欧美日韩一区二区三区在线| av超薄肉色丝袜交足视频| 日日摸夜夜添夜夜添小说| 99re在线观看精品视频| 99香蕉大伊视频| 欧美性长视频在线观看| 天堂动漫精品| 校园春色视频在线观看| 亚洲一区二区三区色噜噜| 国产伦人伦偷精品视频| 国内精品久久久久精免费| 中文字幕人成人乱码亚洲影| 香蕉国产在线看| 亚洲自偷自拍图片 自拍| 欧美老熟妇乱子伦牲交| 老司机午夜十八禁免费视频| 亚洲欧美精品综合一区二区三区| videosex国产| 亚洲av电影不卡..在线观看| 国产精品综合久久久久久久免费 | 亚洲免费av在线视频| 妹子高潮喷水视频| 免费女性裸体啪啪无遮挡网站| 看免费av毛片| 日韩精品免费视频一区二区三区| 国产麻豆成人av免费视频| 亚洲av五月六月丁香网| 久久久精品欧美日韩精品| 最近最新中文字幕大全电影3 | 一区二区三区国产精品乱码| 欧美午夜高清在线| 日韩一卡2卡3卡4卡2021年| 一级黄色大片毛片| 久久人妻福利社区极品人妻图片| 精品一区二区三区av网在线观看| av欧美777| 免费不卡黄色视频| 在线观看66精品国产| 国内毛片毛片毛片毛片毛片| 50天的宝宝边吃奶边哭怎么回事| 男人舔女人下体高潮全视频| 久99久视频精品免费| 一边摸一边抽搐一进一出视频| 亚洲熟妇熟女久久| 高清在线国产一区| 欧美在线黄色| 国产亚洲欧美98| 国产欧美日韩一区二区三区在线| 久久精品91无色码中文字幕| 91在线观看av| 天堂动漫精品| 国产日韩一区二区三区精品不卡| 丁香欧美五月| 人人妻人人澡人人看| 久久国产精品影院| 免费观看人在逋| 黄片小视频在线播放| 女人被狂操c到高潮| 中文亚洲av片在线观看爽| 久久精品国产亚洲av高清一级| 波多野结衣巨乳人妻| 国产欧美日韩精品亚洲av| 亚洲欧洲精品一区二区精品久久久| 国产三级在线视频| 给我免费播放毛片高清在线观看| 欧美激情久久久久久爽电影 | 久久热在线av| 欧美av亚洲av综合av国产av| 可以在线观看的亚洲视频| 午夜两性在线视频| 一个人观看的视频www高清免费观看 | 国产一区二区在线av高清观看| 日本三级黄在线观看| 精品国产一区二区三区四区第35| aaaaa片日本免费| 国产精品综合久久久久久久免费 | 亚洲avbb在线观看| 好男人电影高清在线观看| 91国产中文字幕| 亚洲欧美日韩另类电影网站| 一夜夜www| 中文字幕高清在线视频| 十八禁网站免费在线| 国产精品久久电影中文字幕| 日本五十路高清| 久久久精品国产亚洲av高清涩受| 国产午夜福利久久久久久| a级毛片在线看网站| 久久精品亚洲熟妇少妇任你| 天天添夜夜摸| 成人精品一区二区免费| 啦啦啦 在线观看视频| 欧美一级a爱片免费观看看 | 久久国产精品人妻蜜桃| 两个人免费观看高清视频| 亚洲第一欧美日韩一区二区三区| 纯流量卡能插随身wifi吗| 亚洲成av人片免费观看| 精品少妇一区二区三区视频日本电影| 国产成人精品无人区| av电影中文网址| av在线天堂中文字幕| 亚洲国产欧美网| 亚洲成av片中文字幕在线观看| 国产片内射在线| 男人舔女人下体高潮全视频| 色av中文字幕| 亚洲美女黄片视频| 三级毛片av免费| 国产一区二区三区视频了| 一边摸一边抽搐一进一小说| 国产乱人伦免费视频| 免费人成视频x8x8入口观看| 欧美国产精品va在线观看不卡| 国产一区二区三区综合在线观看| 最近最新中文字幕大全电影3 | 两性午夜刺激爽爽歪歪视频在线观看 | 黄色片一级片一级黄色片| 妹子高潮喷水视频| 亚洲人成电影观看| 久久国产精品影院| 99riav亚洲国产免费| 国产精品一区二区精品视频观看| 亚洲七黄色美女视频| 好看av亚洲va欧美ⅴa在| 国产精品日韩av在线免费观看 | 色综合亚洲欧美另类图片| 多毛熟女@视频| 日本 av在线| 亚洲欧美精品综合久久99| 午夜福利欧美成人| 亚洲av电影不卡..在线观看| 一a级毛片在线观看| 精品日产1卡2卡| 99re在线观看精品视频| 国产97色在线日韩免费| 伊人久久大香线蕉亚洲五| 亚洲欧美激情综合另类| 午夜影院日韩av| 欧美黄色淫秽网站| 波多野结衣av一区二区av| 精品无人区乱码1区二区| 久久人人97超碰香蕉20202| 久久久国产成人免费| 一进一出抽搐gif免费好疼| 亚洲人成电影免费在线| 一进一出抽搐gif免费好疼| 波多野结衣一区麻豆| 999精品在线视频| 久久性视频一级片| 亚洲精品在线美女| 波多野结衣一区麻豆| 女人被躁到高潮嗷嗷叫费观| tocl精华| 黄色 视频免费看| 可以在线观看的亚洲视频| 欧美+亚洲+日韩+国产| aaaaa片日本免费| 亚洲成人精品中文字幕电影| 成熟少妇高潮喷水视频| 欧美黑人精品巨大| 黄色丝袜av网址大全| 久久精品国产清高在天天线| 日本在线视频免费播放| 一进一出好大好爽视频| 国产aⅴ精品一区二区三区波| 日本精品一区二区三区蜜桃| 精品一区二区三区av网在线观看| АⅤ资源中文在线天堂| 亚洲av成人一区二区三| 亚洲狠狠婷婷综合久久图片| 精品卡一卡二卡四卡免费| 免费看美女性在线毛片视频| 欧美绝顶高潮抽搐喷水| 久久亚洲真实| 亚洲欧美激情在线| 国产欧美日韩一区二区精品| 欧美+亚洲+日韩+国产| 少妇的丰满在线观看|