Rapid urbanization and economic development have led to the development of heavy industry and structural re-equalization in mainland China. This has resulted in scattered and disorderly layouts becoming prominent in the region. Furthermore, economic development has exacerbated pressures on regional resources and the environment and has threatened sustainable and coordinated development in the region. The NASA Land Science Investigator Processing System (Land-SIPS) Visible Infrared Imaging Radiometer (VIIRS) 375-m active fire product (VNP14IMG) was selected from the Fire Information for Resource Management System (FIRMS) to study the spatiotemporal patterns of heavy industry development. Furthermore, we employed an improved adaptive K-means algorithm to realize the spatial segmentation of long-order VNP14IMG and constructed heat source objects. Lastly, we used a threshold recognition model to identify heavy industry objects from normal heat source objects. Results suggest that the method is an accurate and effective way to monitor heat sources generated from heavy industry. Moreover, some conclusions about heavy industrial heat source distribution in mainland China at different scales were obtained. Those can be beneficial for policy-makers and heavy industry regulation.
Abstract:Remote sensing (RS) images play a significant role in disaster emergency response. Web2.0 changes the way data are created, making it possible for the public to participate in scientific issues. In this paper, an experiment is designed to evaluate the reliability of crowdsourcing buildings collapse assessment in the early time after an earthquake based on aerial remote sensing image. The procedure of RS data pre-processing and crowdsourcing data collection is presented. A probabilistic model including maximum likelihood estimation (MLE), Bayes' theorem and expectation-maximization (EM) algorithm are applied to quantitatively estimate the individual error-rate and "ground truth" according to multiple participants' assessment results. An experimental area of Yushu earthquake is provided to present the results contributed by participants. Following the results, some discussion is provided regarding accuracy and variation among participants. The features of buildings labeled as the same damage type are found highly consistent. This suggests that the building damage assessment contributed by crowdsourcing can be treated as reliable samples. This study shows potential for a rapid building collapse assessment through crowdsourcing and quantitatively inferring "ground truth" according to crowdsourcing data in the early time after the earthquake based on aerial remote sensing image.
Considering the classification of high spatial resolution remote sensing imagery, this paper presents a novel classification method for such imagery using deep neural networks. Deep learning methods, such as a fully convolutional network (FCN) model, achieve state-of-the-art performance in natural image semantic segmentation when provided with large-scale datasets and respective labels. To use data efficiently in the training stage, we first pre-segment training images and their labels into small patches as supplements of training data using graph-based segmentation and the selective search method. Subsequently, FCN with atrous convolution is used to perform pixel-wise classification. In the testing stage, post-processing with fully connected conditional random fields (CRFs) is used to refine results. Extensive experiments based on the Vaihingen dataset demonstrate that our method performs better than the reference state-of-the-art networks when applied to high-resolution remote sensing imagery classification. accuracy in certain situations. Considering the complexity of land cover that contains vegetation, water, soil and other physical land features, including those created solely by human activities, it is still challenging for object-oriented approaches to improve classification accuracy.Traditionally, classification of remote sensing imagery focused on two steps, i.e., feature extraction methods and supervised learning algorithms. Various methods have been proposed to extract features using hand-crafted descriptors, such as Scale-Invariant Feature Transform (SIFT) [11] and Histogram of Oriented Gradients (HOG) [12]. However, the two steps mentioned above are typically viewed as distinct approaches. Convolutional neural networks (CNNs) combine the two steps into one network that learns contextual features at different scales and computes the score of each class at the end of the network. Another benefit of CNNs is that they can be trained to optimize all weights in the network using an efficient weight update technique such as stochastic gradient descent (SGD) [13] end-to-end. Although CNNs have achieved remarkable results in image categorization, they do not consider pixel-wise semantic classification. In image categorization, the input of a network is an image, while the goal is to predict the correct category of the entire image. FCN [14]-based models have been successfully applied to natural image semantic segmentation and are currently the state-of-the-art for PASCAL VOC datasets [15]. However, FCN models do not take into account relations between pixels and ignore the spatial regularity in the usual pixel-wise classification method. In addition, the vast majority of research focuses on optimization of the network structure without considering efficient use of labeled data that is very important to deep neural networks.To overcome the above challenges, in this paper, we introduce a novel pipeline for pixel-wise classification of high-resolution remote sensing imagery. Our method consists of three parts: ...
Disaster change mapping, which can provide accurate and timely changed information (e.g., damaged buildings, accessibility of road and the shelter sites) for decision makers to guide and support a plan for coordinating emergency rescue, is critical for early disaster rescue. In this paper, we focus on optical remote sensing data to propose an automatic procedure to reduce the impacts of optical data limitations and provide the emergency information in the early phases of a disaster. The procedure utilizes a series of new methods, such as an Optimizable Variational Model (OptVM) for image fusion and a scale-invariant feature transform (SIFT) constraint optical flow method (SIFT-OFM) for image registration, to produce product maps including cloudless backdrop maps and change-detection maps for catastrophic event regions, helping people to be aware of the whole scope of the disaster and assess the distribution and magnitude of damage. These product maps have a rather high accuracy as they are based on high precision preprocessing results in spectral consistency and geometric, which compared with traditional fused and registration methods by visual qualitative or quantitative analysis. The procedure is fully automated without any manual intervention to save response time. It also can be applied to many situations.
Class imbalance is a key issue for the application of deep learning for remote sensing image classification because a model generated by imbalanced samples training has low classification accuracy for minority classes. In this study, an accurate classification approach using the multistage sampling method and deep neural networks was proposed to classify imbalanced data. We first balance samples by multistage sampling to obtain the training sets. Then, a state-of-the-art model is adopted by combining the advantages of atrous spatial pyramid pooling (ASPP) and Encoder-Decoder for pixel-wise classification, which are two different types of fully convolutional networks (FCNs) that can obtain contextual information of multiple levels in the Encoder stage. The details and spatial dimensions of targets are restored using such information during the Decoder stage. We employ four deep learning-based classification algorithms (basic FCN, FCN-8S, ASPP, and Encoder-Decoder with ASPP of our approach) on multistage training sets (original, MUS1, and MUS2) of WorldView-3 images in southeastern Qinghai-Tibet Plateau and GF-2 images in northeastern Beijing for comparison. The experiments show that, compared with existing sets (original, MUS1, and identical) and existing method (cost weighting), the MUS2 training set of multistage sampling significantly enhance the classification performance for minority classes. Our approach shows distinct advantages for imbalanced data.
Abstract:With the rapid development of satellite remote sensing technology, the size of image datasets in many application areas is growing exponentially and the demand for Land-Cover and Land-Use change remote sensing data is growing rapidly. It is thus becoming hard to efficiently and intelligently retrieve the change information that users need from massive image databases. In this paper, content-based image retrieval is successfully applied to change detection, and a content-based remote sensing image change information retrieval model is introduced. First, the construction of a new model framework for change information retrieval from a remote sensing database is described. Then, as the target content cannot be expressed by one kind of feature alone, a multiple-feature, integrated retrieval model is proposed. Thirdly, an experimental prototype system that was set up to demonstrate the validity and practicability of the model is described. The proposed model is a new method of acquiring change detection information from remote sensing imagery and so can reduce the need for image pre-processing and also deal with problems related to seasonal changes, as well as other problems encountered in the field of change detection. Meanwhile, the new model has important implications for improving remote sensing image management and autonomous information retrieval. The experiment results obtained using a Landsat data set show that the use of the new model can produce promising results. A coverage rate and mean average precision of 71% and 89%, respectively, were achieved for the top 20 returned pairs of images.
Abstract:With the rapid development of satellite remote sensing technology, the volume of image datasets in many application areas is growing exponentially and the demand for Land-Cover and Land-Use change remote sensing data is growing rapidly. It is thus becoming hard to efficiently and intelligently retrieve the change information that users need from massive image databases. In this paper, content-based image retrieval is successfully applied to change detection and a content-based remote sensing image change information retrieval model is introduced. First, the construction of a new model framework for change information retrieval in a remote sensing database is described. Then, as the target content cannot be expressed by one kind of feature alone, a multiple-feature integrated retrieval model is proposed. Thirdly, an experimental prototype system that was set up to demonstrate the validity and practicability of the model is described. The proposed model is a new method of acquiring change detection information from remote sensing imagery and so can reduce the need for image pre-processing, deal with problems related toseasonal changes as well as other problems encountered in the field of change detection. Meanwhile, the new model has important implications for improving remote sensing image management and autonomous information retrieval.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.