This study proposes and evaluates five deep fully convolutional networks (FCNs) for the semantic segmentation of a single tree species: SegNet, U-Net, FC-DenseNet, and two DeepLabv3+ variants. The ...performance of the FCN designs is evaluated experimentally in terms of classification accuracy and computational load. We also verify the benefits of fully connected conditional random fields (CRFs) as a post-processing step to improve the segmentation maps. The analysis is conducted on a set of images captured by an RGB camera aboard a UAV flying over an urban area. The dataset also contains a mask that indicates the occurrence of an endangered species called
Vogel, also known as cumbaru, taken as the species to be identified. The experimental analysis shows the effectiveness of each design and reports average overall accuracy ranging from 88.9% to 96.7%, an F1-score between 87.0% and 96.1%, and IoU from 77.1% to 92.5%. We also realize that CRF consistently improves the performance, but at a high computational cost.
Image‐based gauging stations can allow for significant densification of monitoring networks of river water stages. However, thus far, most camera gauges do not provide the robustness of accurate ...measurements due to the varying appearance of water in the stream throughout the year. We introduce an approach that allows for automatic and reliable water stage measurement combining deep learning and photogrammetric techniques. First, a convolutional neural network (CNN), a class of deep learning, is applied to the segmentation (i.e., pixel classification) of water in images. The CNNs SegNet and fully convolutional network (FCN) are associated with a transfer learning strategy to segment water on images acquired by a Raspberry Pi camera. Errors of water segmentation with the two CNNs are lower than 3%. Second, the image information is transformed into metric water stage values by intersecting the extracted water contour, generated using the segmentation results, with a 3D model reconstructed with structure‐from‐motion (SfM) photogrammetry. The highest correlations between a reference gauge and the image‐based approaches reached 0.93, and average deviations were lower than 4 cm. Our approach allows for the densification of river monitoring networks based on camera gauges, providing accurate water stage measurements.
Key Points
Development of a camera gauge to measure the water stage from single images using convolutional neural network (CNN)
Combining automatic/robust water area detection with high‐resolution 3D data for stage retrieval in challenging environmental conditions
CNN based water stage measurements reveal accuracy potential at cm‐range and up to 93% agreement with traditional reference gauge
A review on deep learning in UAV remote sensing Osco, Lucas Prado; Marcato Junior, José; Marques Ramos, Ana Paula ...
International journal of applied earth observation and geoinformation,
October 2021, 2021-10-00, 2021-10-01, Letnik:
102
Journal Article
Recenzirano
Odprti dostop
•Combining deep learning and UAV-based data is an emerging trend in remote sensing.•Most articles published rely on CNN-based methods.•Future perspectives in UAV-based data processing still have much ...to cover.
Deep Neural Networks (DNNs) learn representation from data with an impressive capability, and brought important breakthroughs for processing images, time-series, natural language, audio, video, and many others. In the remote sensing field, surveys and literature revisions specifically involving DNNs algorithms’ applications have been conducted in an attempt to summarize the amount of information produced in its subfields. Recently, Unmanned Aerial Vehicle (UAV)-based applications have dominated aerial sensing research. However, a literature revision that combines both “deep learning” and “UAV remote sensing” thematics has not yet been conducted. The motivation for our work was to present a comprehensive review of the fundamentals of Deep Learning (DL) applied in UAV-based imagery. We focused mainly on describing the classification and regression techniques used in recent applications with UAV-acquired data. For that, a total of 232 papers published in international scientific journal databases was examined. We gathered the published materials and evaluated their characteristics regarding the application, sensor, and technique used. We discuss how DL presents promising results and has the potential for processing tasks associated with UAV-based image data. Lastly, we project future perspectives, commentating on prominent DL paths to be explored in the UAV remote sensing field. This revision consisting of an approach to introduce, commentate, and summarize the state-of-the-art in UAV-based image applications with DNNs algorithms in diverse subfields of remote sensing, grouping it in the environmental, urban, and agricultural contexts.
The traditional method of measuring nitrogen content in plants is a time-consuming and labor-intensive task. Spectral vegetation indices extracted from unmanned aerial vehicle (UAV) images and ...machine learning algorithms have been proved effective in assisting nutritional analysis in plants. Still, this analysis has not considered the combination of spectral indices and machine learning algorithms to predict nitrogen in tree-canopy structures. This paper proposes a new framework to infer the nitrogen content in citrus-tree at a canopy-level using spectral vegetation indices processed with the random forest algorithm. A total of 33 spectral indices were estimated from multispectral images acquired with a UAV-based sensor. Leaf samples were gathered from different planting-fields and the leaf nitrogen content (LNC) was measured in the laboratory, and later converted into the canopy nitrogen content (CNC). To evaluate the robustness of the proposed framework, we compared it with other machine learning algorithms. We used 33,600 citrus trees to evaluate the performance of the machine learning models. The random forest algorithm had higher performance in predicting CNC than all models tested, reaching an R2 of 0.90, MAE of 0.341 g·kg−1 and MSE of 0.307 g·kg−1. We demonstrated that our approach is able to reduce the need for chemical analysis of the leaf tissue and optimizes citrus orchard CNC monitoring.
Monitoring biomass of forages in experimental plots and livestock farms is a time-consuming, expensive, and biased task. Thus, non-destructive, accurate, precise, and quick phenotyping strategies for ...biomass yield are needed. To promote high-throughput phenotyping in forages, we propose and evaluate the use of deep learning-based methods and UAV (Unmanned Aerial Vehicle)-based RGB images to estimate the value of biomass yield by different genotypes of the forage grass species
Jacq. Experiments were conducted in the Brazilian Cerrado with 110 genotypes with three replications, totaling 330 plots. Two regression models based on Convolutional Neural Networks (CNNs) named AlexNet and ResNet18 were evaluated, and compared to VGGNet-adopted in previous work in the same thematic for other grass species. The predictions returned by the models reached a correlation of 0.88 and a mean absolute error of 12.98% using AlexNet considering pre-training and data augmentation. This proposal may contribute to forage biomass estimation in breeding populations and livestock areas, as well as to reduce the labor in the field.
Urban forests are an important part of any city, given that they provide several environmental benefits, such as improving urban drainage, climate regulation, public health, biodiversity, and others. ...However, tree detection in cities is challenging, given the irregular shape, size, occlusion, and complexity of urban areas. With the advance of environmental technologies, deep learning segmentation mapping methods can map urban forests accurately. We applied a region-based CNN object instance segmentation algorithm for the semantic segmentation of tree canopies in urban environments based on aerial RGB imagery. To the best of our knowledge, no study investigated the performance of deep learning-based methods for segmentation tasks inside the Cerrado biome, specifically for urban tree segmentation. Five state-of-the-art architectures were evaluated, namely: Fully Convolutional Network; U-Net; SegNet; Dynamic Dilated Convolution Network and DeepLabV3+. The experimental analysis showed the effectiveness of these methods reporting results such as pixel accuracy of 96,35%, an average accuracy of 91.25%, F1-score of 91.40%, Kappa of 82.80% and IoU of 73.89%. We also determined the inference time needed per area, and the deep learning methods investigated after the training proved to be suitable to solve this task, providing fast and effective solutions with inference time varying from 0.042 to 0.153 minutes per hectare. We conclude that the semantic segmentation of trees inside urban environments is highly achievable with deep neural networks. This information could be of high importance to decision-making and may contribute to the management of urban systems. It should be also important to mention that the dataset used in this work is available on our website.
Deep neural networks are currently the focus of many remote sensing approaches related to forest management. Although they return satisfactory results in most tasks, some challenges related to ...hyperspectral data remain, like the curse of data dimensionality. In forested areas, another common problem is the highly-dense distribution of trees. In this paper, we propose a novel deep learning approach for hyperspectral imagery to identify single-tree species in highly-dense areas. We evaluated images with 25 spectral bands ranging from 506 to 820 nm taken over a semideciduous forest of the Brazilian Atlantic biome. We included in our network’s architecture a band combination selection phase. This phase learns from multiple combinations between bands which contributed the most for the tree identification task. This is followed by a feature map extraction and a multi-stage model refinement of the confidence map to produce accurate results of a highly-dense target. Our method returned an f-measure, precision and recall values of 0.959, 0.973, and 0.945, respectively. The results were superior when compared with a principal component analysis (PCA) approach. Compared to other learning methods, ours estimate a combination of hyperspectral bands that most contribute to the mentioned task within the network’s architecture. With this, the proposed method achieved state-of-the-art performance for detecting and geolocating individual tree-species in UAV-based hyperspectral images in a complex forest.
An earthquake-induced landslide (EQIL) is a rapidly changing process occurring at the Earth’s surface that is strongly controlled by the earthquake in question and predisposing conditions. Predicting ...locations prone to EQILs on a large scale is significant for managing rescue operations and disaster mitigation. We propose a deep learning framework while considering the source area feature of EQIL to model the complex relationship and enhance spatial prediction accuracy. Initially, we used high-resolution remote sensing images and a digital elevation model (DEM) to extract the source area of an EQIL. Then, 14 controlling factors were input to a stacked autoencoder (SAE) to search for robust features by sparse optimization, and the classifier took advantage of high-level abstract features to identify the EQIL spatially. Finally, the EQIL inventory collected from the Wenchuan earthquake was used to validate the proposed model. The results show that the proposed method significantly outperformed conventional methods, achieving an overall accuracy (OA) of 91.88%, while logistic regression (LR), support vector machine (SVM), and random forest (RF) achieved 80.75%, 82.22%, and 84.16%, respectively. Meanwhile, this study reveals that shallow machine learning models only take advantage of significant factors for EQIL prediction, but deep learning models can extract more effective information related to EQIL distribution from low-value density data, which is why its prediction accuracy is growing with increasing input factors. There is hope that new knowledge of EQILs can be represented by high-level abstract features extracted by hidden layers of the deep learning model, which are typically acquired by statistical methods.
Road extraction in remote sensing data: A survey Chen, Ziyi; Deng, Liai; Luo, Yuhua ...
International journal of applied earth observation and geoinformation,
August 2022, 2022-08-00, 2022-08-01, Letnik:
112
Journal Article
Recenzirano
Odprti dostop
•This review covers a wider perspective in terms of both 2D remote sensing images and 3D point clouds.•This review provides a detail survey on 2D and 3D remote sensing datasets used for road ...extraction.•This review presents a detail analysis of challenges and future trends of road extraction from remote sensing data.
Automated extraction of roads from remotely sensed data come forth various usages ranging from digital twins for smart cities, intelligent transportation, urban planning, autonomous driving, to emergency management. Many studies have focused on promoting the progress of methods for automated road extraction from aerial and satellite optical images, synthetic aperture radar (SAR) images, and LiDAR point clouds. In the past 10 years, no a more comprehensive survey on this topic could be found in literature. This paper attempts to provide a comprehensive survey on road extraction methods that use 2D earth observing images and 3D LiDAR point clouds. In this review, we first present a tree-structure that separate the literature into 2D and 3D. Then, further methodologies level classification is demonstrated both in 2D and 3D. In 2D and 3D, we introduce and analyze the literature published in the last ten years. Except for the methodologies, we also review the aspects of data commonly used. Finally, this paper explores the existing challenges and future trends.
We assessed the performance of Convolutional Neural Network (CNN)-based approaches using mobile phone images to estimate regrowth density in tropical forages. We generated a dataset composed of 1124 ...labeled images with 2 mobile phones 7 days after the harvest of the forage plants. Six architectures were evaluated, including AlexNet, ResNet (18, 34, and 50 layers), ResNeXt101, and DarkNet. The best regression model showed a mean absolute error of 7.70 and a correlation of 0.89. Our findings suggest that our proposal using deep learning on mobile phone images can successfully be used to estimate regrowth density in forages.