Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
Olive trees, which are planted widely in China, are economically significant. Timely and accurate acquisition of olive tree crown information is vital in monitoring olive tree growth and accurately predicting its fruit yield. The advent of unmanned aerial vehicles (UAVs) and deep learning (DL) provides an opportunity for rapid monitoring parameters of the olive tree crown. In this study, we propose a method of automatically extracting olive crown information (crown number and area of olive tree), combining visible-light images captured by consumer UAV and a new deep learning model, U2-Net, with a deeply nested structure. Firstly, a data set of an olive tree crown (OTC) images was constructed, which was further processed by the ESRGAN model to enhance the image resolution and was augmented (geometric transformation and spectral transformation) to enlarge the data set to increase the generalization ability of the model. Secondly, four typical subareas (A–D) in the study area were selected to evaluate the performance of the U2-Net model in olive crown extraction in different scenarios, and the U2-Net model was compared with three current mainstream deep learning models (i.e., HRNet, U-Net, and DeepLabv3+) in remote sensing image segmentation effect. The results showed that the U2-Net model achieved high accuracy in the extraction of tree crown numbers in the four subareas with a mean of intersection over union (IoU), overall accuracy (OA), and F1-Score of 92.27%, 95.19%, and 95.95%, respectively. Compared with the other three models, the IoU, OA, and F1-Score of the U2-Net model increased by 14.03–23.97 percentage points, 7.57–12.85 percentage points, and 8.15–14.78 percentage points, respectively. In addition, the U2-Net model had a high consistency between the predicted and measured area of the olive crown, and compared with the other three deep learning models, it had a lower error rate with a root mean squared error (RMSE) of 4.78, magnitude of relative error (MRE) of 14.27%, and a coefficient of determination (R2) higher than 0.93 in all four subareas, suggesting that the U2-Net model extracted the best crown profile integrity and was most consistent with the actual situation. This study indicates that the method combining UVA RGB images with the U2-Net model can provide a highly accurate and robust extraction result for olive tree crowns and is helpful in the dynamic monitoring and management of orchard trees.
Olive trees, which are planted widely in China, are economically significant. Timely and accurate acquisition of olive tree crown information is vital in monitoring olive tree growth and accurately predicting its fruit yield. The advent of unmanned aerial vehicles (UAVs) and deep learning (DL) provides an opportunity for rapid monitoring parameters of the olive tree crown. In this study, we propose a method of automatically extracting olive crown information (crown number and area of olive tree), combining visible-light images captured by consumer UAV and a new deep learning model, U2-Net, with a deeply nested structure. Firstly, a data set of an olive tree crown (OTC) images was constructed, which was further processed by the ESRGAN model to enhance the image resolution and was augmented (geometric transformation and spectral transformation) to enlarge the data set to increase the generalization ability of the model. Secondly, four typical subareas (A–D) in the study area were selected to evaluate the performance of the U2-Net model in olive crown extraction in different scenarios, and the U2-Net model was compared with three current mainstream deep learning models (i.e., HRNet, U-Net, and DeepLabv3+) in remote sensing image segmentation effect. The results showed that the U2-Net model achieved high accuracy in the extraction of tree crown numbers in the four subareas with a mean of intersection over union (IoU), overall accuracy (OA), and F1-Score of 92.27%, 95.19%, and 95.95%, respectively. Compared with the other three models, the IoU, OA, and F1-Score of the U2-Net model increased by 14.03–23.97 percentage points, 7.57–12.85 percentage points, and 8.15–14.78 percentage points, respectively. In addition, the U2-Net model had a high consistency between the predicted and measured area of the olive crown, and compared with the other three deep learning models, it had a lower error rate with a root mean squared error (RMSE) of 4.78, magnitude of relative error (MRE) of 14.27%, and a coefficient of determination (R2) higher than 0.93 in all four subareas, suggesting that the U2-Net model extracted the best crown profile integrity and was most consistent with the actual situation. This study indicates that the method combining UVA RGB images with the U2-Net model can provide a highly accurate and robust extraction result for olive tree crowns and is helpful in the dynamic monitoring and management of orchard trees.
The determination of key phenological growth stages of banana plantations, such as flower emergence and plant establishment, is difficult due to the asynchronous growth habit of banana plants. Identifying phenological events assists growers in determining plant maturity, and harvest timing and guides the application of time-specific crop inputs. Currently, phenological monitoring requires repeated manual observations of individual plants’ growth stages, which is highly laborious, time-inefficient, and requires the handling and integration of large field-based data sets. The ability of growers to accurately forecast yield is also compounded by the asynchronous growth of banana plants. Satellite remote sensing has proved effective in monitoring spatial and temporal crop phenology in many broadacre crops. However, for banana crops, very high spatial and temporal resolution imagery is required to enable individual plant level monitoring. Unoccupied aerial vehicle (UAV)-based sensing technologies provide a cost-effective solution, with the potential to derive information on health, yield, and growth in a timely, consistent, and quantifiable manner. Our research explores the ability of UAV-derived data to track temporal phenological changes of individual banana plants from follower establishment to harvest. Individual plant crowns were delineated using object-based image analysis, with calculations of canopy height and canopy area producing strong correlations against corresponding ground-based measures of these parameters (R2 of 0.77 and 0.69 respectively). A temporal profile of canopy reflectance and plant morphology for 15 selected banana plants were derived from UAV-captured multispectral data over 21 UAV campaigns. The temporal profile was validated against ground-based determinations of key phenological growth stages. Derived measures of minimum plant height provided the strongest correlations to plant establishment and harvest, whilst interpolated maxima of normalised difference vegetation index (NDVI) best indicated flower emergence. For pre-harvest yield forecasting, the Enhanced Vegetation Index 2 provided the strongest relationship (R2 = 0.77) from imagery captured near flower emergence. These findings demonstrate that UAV-based multitemporal crop monitoring of individual banana plants can be used to determine key growing stages of banana plants and offer pre-harvest yield forecasts.
When it comes to forest management and protection, knowledge is key. Therefore, forest mapping is crucial to obtain the required knowledge towards profitable resource exploitation and increased resilience against wildfires. Within this context, this paper presents a literature review on tree classification and segmentation using data acquired by unmanned aerial vehicles, with special focus on the last decade (2013–2023). The latest research trends in this field are presented and analyzed in two main vectors, namely: (1) data, where used sensors and data structures are resumed; and (2) methods, where remote sensing and data analysis methods are described, with particular focus on machine learning approaches. The study and review methodology filtered 979 papers, which were then screened, resulting in the 144 works included in this paper. These are systematically analyzed and organized by year, keywords, purpose, sensors, and methods used, easily allowing the readers to have a wide, but at the same time detailed, view of the latest trends in automatic tree classification and segmentation using unmanned aerial vehicles. This review shows that image processing and machine learning techniques applied to forestry and segmentation and classification tasks are focused on improving the accuracy and interpretability of the results by using multi-modal data, 3D information, and AI methods. Most works use RGB or multispectral cameras, or LiDAR scanners, individually. Classification is mostly carried out using supervised methods, while segmentation mostly uses unsupervised machine learning techniques.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.