Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
2,472
result(s) for
"tree crown"
Sort by:
Multi-Species Individual Tree Segmentation and Identification Based on Improved Mask R-CNN and UAV Imagery in Mixed Forests
2022
High-resolution UAV imagery paired with a convolutional neural network approach offers significant advantages in accurately measuring forestry ecosystems. Despite numerous studies existing for individual tree crown delineation, species classification, and quantity detection, the comprehensive situation in performing the above tasks simultaneously has rarely been explored, especially in mixed forests. In this study, we propose a new method for individual tree segmentation and identification based on the improved Mask R-CNN. For the optimized network, the fusion type in the feature pyramid network is modified from down-top to top-down to shorten the feature acquisition path among the different levels. Meanwhile, a boundary-weighted loss module is introduced to the cross-entropy loss function Lmask to refine the target loss. All geometric parameters (contour, the center of gravity and area) associated with canopies ultimately are extracted from the mask by a boundary segmentation algorithm. The results showed that F1-score and mAP for coniferous species were higher than 90%, and that of broadleaf species were located between 75–85.44%. The producer’s accuracy of coniferous forests was distributed between 0.8–0.95 and that of broadleaf ranged in 0.87–0.93; user’s accuracy of coniferous was distributed between 0.81–0.84 and that of broadleaf ranged in 0.71–0.76. The total number of trees predicted was 50,041 for the entire study area, with an overall error of 5.11%. The method under study is compared with other networks including U-net and YOLOv3. Results in this study show that the improved Mask R-CNN has more advantages in broadleaf canopy segmentation and number detection.
Journal Article
Tree Crown Delineation Algorithm Based on a Convolutional Neural Network
by
P. Ferreira, Matheus
,
F. de Campos Velho, Haroldo
,
Peripato, Vinícius
in
algorithms
,
biomass
,
deep learning
2020
Tropical forests concentrate the largest diversity of species on the planet and play a key role in maintaining environmental processes. Due to the importance of those forests, there is growing interest in mapping their components and getting information at an individual tree level to conduct reliable satellite-based forest inventory for biomass and species distribution qualification. Individual tree crown information could be manually gathered from high resolution satellite images; however, to achieve this task at large-scale, an algorithm to identify and delineate each tree crown individually, with high accuracy, is a prerequisite. In this study, we propose the application of a convolutional neural network—Mask R-CNN algorithm—to perform the tree crown detection and delineation. The algorithm uses very high-resolution satellite images from tropical forests. The results obtained are promising—the R e c a l l , P r e c i s i o n , and F 1 score values obtained were were 0.81 , 0.91 , and 0.86 , respectively. In the study site, the total of tree crowns delineated was 59,062 . These results suggest that this algorithm can be used to assist the planning and conduction of forest inventories. As the algorithm is based on a Deep Learning approach, it can be systematically trained and used for other regions.
Journal Article
Linking crown structure with tree ring pattern: methodological considerations and proof of concept
by
Pretzsch, Hans
,
Hilmers, Torben
,
Schmied, Gerhard
in
Diameters
,
Ecosystem management
,
Ecosystem services
2022
Key messageStructural characteristics of tree crowns obtained by TLidar scanning can be used for estimating the course of the stem diameter growth in the past.To improve human well-being through sustainable management of ecosystems, particular attention is given to the structures, functions, and services of forest trees and stands. The classical timber provision has become only one of many other forest ecosystem services. At the same time, the methods of ecosystem observation, analysis, and modelling have enormously improved. Here, we fathomed the information potential of the tree crown structure. Our overarching hypothesis was that the crown structure reflects essential characteristics of the tree ring pattern. The empirical part of this study was based on sample trees from the combined spacing-thinning trial in Norway spruce (Picea abies [L.] Karst.) Fürstenfeldbruck 612 in Southern Germany. First, we showed that the external characteristics of tree crowns and the internal stem structure are functionally linked. Second, we derived metrics for the tree ring pattern and crown shape, and found especially close relationships between the level and bending of the growth curve and the size and stereometric shape of the crown. Third, we investigated how the derived statistical relationships between tree ring pattern and crown structure can be applied to derive the course of tree growth from the crown structure. We showed how measures such as size and variability of the crown could be used to estimate the course of diameter growth. Finally, we showed that the revealed link could be used to assess past and future growth and life expectancy of trees. These findings can be used to monitor the stress defence potential, resistance, and resilience of trees.
Journal Article
Individual Tree Crown Segmentation Directly from UAV-Borne LiDAR Data Using the PointNet of Deep Learning
2021
Accurate individual tree crown (ITC) segmentation from scanned point clouds is a fundamental task in forest biomass monitoring and forest ecology management. Light detection and ranging (LiDAR) as a mainstream tool for forest survey is advancing the pattern of forest data acquisition. In this study, we performed a novel deep learning framework directly processing the forest point clouds belonging to the four forest types (i.e., the nursery base, the monastery garden, the mixed forest, and the defoliated forest) to realize the ITC segmentation. The specific steps of our approach were as follows: first, a voxelization strategy was conducted to subdivide the collected point clouds with various tree species from various forest types into many voxels. These voxels containing point clouds were taken as training samples for the PointNet deep learning framework to identify the tree crowns at the voxel scale. Second, based on the initial segmentation results, we used the height-related gradient information to accurately depict the boundaries of each tree crown. Meanwhile, the retrieved tree crown breadths of individual trees were compared with field measurements to verify the effectiveness of our approach. Among the four forest types, our results revealed the best performance for the nursery base (tree crown detection rate r = 0.90; crown breadth estimation R2 > 0.94 and root mean squared error (RMSE) < 0.2m). A sound performance was also achieved for the monastery garden and mixed forest, which had complex forest structures, complicated intersections of branches and different building types, with r = 0.85, R2 > 0.88 and RMSE < 0.6 m for the monastery garden and r = 0.80, R2 > 0.85 and RMSE < 0.8 m for the mixed forest. For the fourth forest plot type with the distribution of crown defoliation across the woodland, we achieved the performance with r = 0.82, R2 > 0.79 and RMSE < 0.7 m. Our method presents a robust framework inspired by the deep learning technology and computer graphics theory that solves the ITC segmentation problem and retrieves forest parameters under various forest conditions.
Journal Article
A New Individual Tree Species Recognition Method Based on a Convolutional Neural Network and High-Spatial Resolution Remote Sensing Imagery
by
Wang, Huan
,
Yan, Shijie
,
Jing, Linhai
in
convolutional neural network
,
data collection
,
forests
2021
Tree species surveys are crucial to forest resource management and can provide references for forest protection policy making. The traditional tree species survey in the field is labor-intensive and time-consuming, supporting the practical significance of remote sensing. The availability of high-resolution satellite remote sensing data enable individual tree species (ITS) recognition at low cost. In this study, the potential of the combination of such images and a convolutional neural network (CNN) to recognize ITS was explored. Firstly, individual tree crowns were delineated from a high-spatial resolution WorldView-3 (WV3) image and manually labeled as different tree species. Next, a dataset of the image subsets of the labeled individual tree crowns was built, and several CNN models were trained based on the dataset for ITS recognition. The models were then applied to the WV3 image. The results show that the distribution maps of six ITS offered an overall accuracy of 82.7% and a kappa coefficient of 0.79 based on the modified GoogLeNet, which used the multi-scale convolution kernel to extract features of the tree crown samples and was modified for small-scale samples. The ITS recognition method proposed in this study, with multi-scale individual tree crown delineation, avoids artificial tree crown delineation. Compared with the random forest (RF) and support vector machine (SVM) approaches, this method can automatically extract features and outperform RF and SVM in the classification of six tree species.
Journal Article
Tree Crown Detection and Delineation in a Temperate Deciduous Forest from UAV RGB Imagery Using Deep Learning Approaches: Effects of Spatial Resolution and Species Characteristics
2023
The automatic detection of tree crowns and estimation of crown areas from remotely sensed information offer a quick approach for grasping the dynamics of forest ecosystems and are of great significance for both biodiversity and ecosystem conservation. Among various types of remote sensing data, unmanned aerial vehicle (UAV)-acquired RGB imagery has been increasingly used for tree crown detection and crown area estimation; the method has efficient advantages and relies heavily on deep learning models. However, the approach has not been thoroughly investigated in deciduous forests with complex crown structures. In this study, we evaluated two widely used, deep-learning-based tree crown detection and delineation approaches (DeepForest and Detectree2) to assess their potential for detecting tree crowns from UAV-acquired RGB imagery in an alpine, temperate deciduous forest with a complicated species composition. A total of 499 digitized crowns, including four dominant species, with corresponding, accurate inventory data in a 1.5 ha study plot were treated as training and validation datasets. We attempted to identify an effective model to delineate tree crowns and to explore the effects of the spatial resolution on the detection performance, as well as the extracted tree crown areas, with a detailed field inventory. The results show that the two deep-learning-based models, of which Detectree2 (F1 score: 0.57) outperformed DeepForest (F1 score: 0.52), could both be transferred to predict tree crowns successfully. However, the spatial resolution had an obvious effect on the estimation accuracy of tree crown detection, especially when the resolution was greater than 0.1 m. Furthermore, Dectree2 could estimate tree crown areas accurately, highlighting its potential and robustness for tree detection and delineation. In addition, the performance of tree crown detection varied among different species. These results indicate that the evaluated approaches could efficiently delineate individual tree crowns in high-resolution optical images, while demonstrating the applicability of Detectree2, and, thus, have the potential to offer transferable strategies that can be applied to other forest ecosystems.
Journal Article
Extraction of Olive Crown Based on UAV Visible Images and the U2-Net Deep Learning Model
2022
Olive trees, which are planted widely in China, are economically significant. Timely and accurate acquisition of olive tree crown information is vital in monitoring olive tree growth and accurately predicting its fruit yield. The advent of unmanned aerial vehicles (UAVs) and deep learning (DL) provides an opportunity for rapid monitoring parameters of the olive tree crown. In this study, we propose a method of automatically extracting olive crown information (crown number and area of olive tree), combining visible-light images captured by consumer UAV and a new deep learning model, U2-Net, with a deeply nested structure. Firstly, a data set of an olive tree crown (OTC) images was constructed, which was further processed by the ESRGAN model to enhance the image resolution and was augmented (geometric transformation and spectral transformation) to enlarge the data set to increase the generalization ability of the model. Secondly, four typical subareas (A–D) in the study area were selected to evaluate the performance of the U2-Net model in olive crown extraction in different scenarios, and the U2-Net model was compared with three current mainstream deep learning models (i.e., HRNet, U-Net, and DeepLabv3+) in remote sensing image segmentation effect. The results showed that the U2-Net model achieved high accuracy in the extraction of tree crown numbers in the four subareas with a mean of intersection over union (IoU), overall accuracy (OA), and F1-Score of 92.27%, 95.19%, and 95.95%, respectively. Compared with the other three models, the IoU, OA, and F1-Score of the U2-Net model increased by 14.03–23.97 percentage points, 7.57–12.85 percentage points, and 8.15–14.78 percentage points, respectively. In addition, the U2-Net model had a high consistency between the predicted and measured area of the olive crown, and compared with the other three deep learning models, it had a lower error rate with a root mean squared error (RMSE) of 4.78, magnitude of relative error (MRE) of 14.27%, and a coefficient of determination (R2) higher than 0.93 in all four subareas, suggesting that the U2-Net model extracted the best crown profile integrity and was most consistent with the actual situation. This study indicates that the method combining UVA RGB images with the U2-Net model can provide a highly accurate and robust extraction result for olive tree crowns and is helpful in the dynamic monitoring and management of orchard trees.
Journal Article
Individual Tree Crown Segmentation of a Larch Plantation Using Airborne Laser Scanning Data Based on Region Growing and Canopy Morphology Features
2020
The detection of individual trees in a larch plantation could improve the management efficiency and production prediction. This study introduced a two-stage individual tree crown (ITC) segmentation method for airborne light detection and ranging (LiDAR) point clouds, focusing on larch plantation forests with different stem densities. The two-stage segmentation method consists of the region growing and morphology segmentation, which combines advantages of the region growing characteristics and the detailed morphology structures of tree crowns. The framework comprises five steps: (1) determination of the initial dominant segments using a region growing algorithm, (2) identification of segments to be redefined based on the 2D hull convex area of each segment, (3) establishment and selection of profiles based on the tree structures, (4) determination of the number of trees using the correlation coefficient of residuals between Gaussian fitting and the tree canopy shape described in each profile, and (5) k-means segmentation to obtain the point cloud of a single tree. The accuracy was evaluated in terms of correct matching, recall, precision, and F-score in eight plots with different stem densities. Results showed that the proposed method significantly increased ITC detections compared with that of using only the region growing algorithm, where the correct matching rate increased from 73.5% to 86.1%, and the recall value increased from 0.78 to 0.89.
Journal Article
Accurate delineation of individual tree crowns in tropical forests from aerial RGB imagery using Mask R‐CNN
by
Jackson, Tobias
,
Department of Plant Sciences (Cambridge, UK) ; University of Cambridge [UK] (CAM)
,
S. H. M. H. received funding from the Centre for Doctoral Training in Application of Artificial Intelligence to the study of Environmental Risks (AI4ER, EP/S022961/1), which is supported by the Engineering and Physical Sciences Research Council (EPSRC). J. G. C. B. was supported by the NERC C-CLEAR doctoral training programme (PDAG/501). T. D. J. and D. A. C. were supported by NERC grant (NE/S010750/1). D. A. C. was supported by the Franklinia Foundation. Data collection in French Guiana was supported by CNES who funded the 2016 hyperspectral, RGB and lidar data over Paracou and Labex CEBA (ANR-10-LABX-25) for contributing financial resource for the field validation of manual crown segmentations. The 2019 data in Paracou and 2020 data in Sabah were funded by NERC (NE/S010750/1). The 2014 Sabah data were also funded by NERC (NE/K016377/1)
in
Aerial photography
,
Aerial surveys
,
Agricultural sciences
2023
Tropical forests are a major component of the global carbon cycle and home to two-thirds of terrestrial species. Upper-canopy trees store the majority of forest carbon and can be vulnerable to drought events and storms. Monitoring their growth and mortality is essential to understanding forest resilience to climate change, but in the context of forest carbon storage, large trees are underrepresented in traditional field surveys, so estimates are poorly constrained. Aerial photographs provide spectral and textural information to discriminate between tree crowns in diverse, complex tropical canopies, potentially opening the door to landscape monitoring of large trees. Here we describe a new deep convolutional neural network method, Detectree2, which builds on the Mask R-CNN computer vision framework to recognize the irregular edges of individual tree crowns from airborne RGB imagery. We trained and evaluated this model with 3797 manually delineated tree crowns at three sites in Malaysian Borneo and one site in French Guiana. As an example application, we combined the delineations with repeat lidar surveys (taken between 3 and 6 years apart) of the four sites to estimate the growth and mortality of upper-canopy trees. Detectree2 delineated 65 000 upper-canopy trees across 14 km2 of aerial images. The skill of the automatic method in delineating unseen test trees was good (F1 score = 0.64) and for the tallest category of trees was excellent (F1 score = 0.74). As predicted from previous field studies, we found that growth rate declined with tree height and tall trees had higher mortality rates than intermediate-size trees. Our approach demonstrates that deep learning methods can automatically segment trees in widely accessible RGB imagery. This tool (provided as an open-source Python package) has many potential applications in forest ecology and conservation, from estimating carbon stocks to monitoring forest phenology and restoration.Python package available to install at https://github.com/PatBall1/ Detectree2.
Journal Article
Individual Rubber Tree Segmentation Based on Ground-Based LiDAR Data and Faster R-CNN of Deep Learning
2019
Rubber trees in southern China are often impacted by natural disturbances that can result in a tilted tree body. Accurate crown segmentation for individual rubber trees from scanned point clouds is an essential prerequisite for accurate tree parameter retrieval. In this paper, three plots of different rubber tree clones, PR107, CATAS 7-20-59, and CATAS 8-7-9, were taken as the study subjects. Through data collection using ground-based mobile light detection and ranging (LiDAR), a voxelisation method based on the scanned tree trunk data was proposed, and deep images (i.e., images normally used for deep learning) were generated through frontal and lateral projection transform of point clouds in each voxel with a length of 8 m and a width of 3 m. These images provided the training and testing samples for the faster region-based convolutional neural network (Faster R-CNN) of deep learning. Consequently, the Faster R-CNN combined with the generated training samples comprising 802 deep images with pre-marked trunk locations was trained to automatically recognize the trunk locations in the testing samples, which comprised 359 deep images. Finally, the point clouds for the lower parts of each trunk were extracted through back-projection transform from the recognized trunk locations in the testing samples and used as the seed points for the region’s growing algorithm to accomplish individual rubber tree crown segmentation. Compared with the visual inspection results, the recognition rate of our method reached 100% for the deep images of the testing samples when the images contained one or two trunks or the trunk information was slightly occluded by leaves. For the complicated cases, i.e., multiple trunks or overlapping trunks in one deep image or a trunk appearing in two adjacent deep images, the recognition accuracy of our method was greater than 90%. Our work represents a new method that combines a deep learning framework with point cloud processing for individual rubber tree crown segmentation based on ground-based mobile LiDAR scanned data.
Journal Article