Big data has now become a strong focus of global interest that is increasingly attracting the attention of academia, industry, government and other organizations. Big data can be situated in the disciplinary area of traditional geospatial data handling theory and methods. The increasing volume and varying format of collected geospatial big data presents challenges in storing, managing, processing, analyzing, visualizing and verifying the quality of data. This has implications for the quality of decisions made with big data. Consequently, this position paper of the International Society for Photogrammetry and Remote Sensing (ISPRS) Technical Commission II (TC II) revisits the existing geospatial data handling methods and theories to determine if they are still capable of handling emerging geospatial big data. Further, the paper synthesises problems, major issues and challenges with current developments as well as recommending what needs to be developed further in the near future.
The landscape in which people live is made up of many features, which are named and have importance for cultural reasons. Prominent among these are the naming of upland features such as mountains, but mountains are an enigmatic phenomenon which do not bear precise and repeatable definition. They have a vague spatial extent, and recent research has modelled such classes as spatial fuzzy sets. We take a specifically multiresolution approach to the definition of the fuzzy set membership of morphometric classes of landscape. We explore this idea with respect to the identification of culturally recognized landscape features of the English Lake District. Discussion focuses on peaks and passes, and the results show that the landscape elements identified in the analysis correspond to well-known landmarks included in a place name database for the area, although many more are found in the analysis than are named in the available database. Further analysis shows that a richer interrogation of the landscape can be achieved with Geographical Information Systems when using this method than using standard approaches. key words fuzzy sets landforms morphometry multi-scale analysis mountains the Lake District
Objectives Decades of empirical research demonstrate that crime is concentrated at a range of spatial scales, including street segments. Further, the degree of clustering at particular geographic units remains noticeably stable and consistent; a finding that Weisburd (Criminology 53:133-157, 2015) has recently termed the 'law of crime concentration at places'. Such findings suggest that the future locations of crime should-to some extent at least-be predictable. To date, methods of forecasting where crime is most likely to next occur have focused either on area-level or grid-based predictions. No studies of which we are aware have developed and tested the accuracy of methods for predicting the future risk of crime at the street segment level. This is surprising given that it is at this level of place that many crimes are committed and policing resources are deployed. Methods Using data for property crimes for a large UK metropolitan police force area, we introduce and calibrate a network-based version of prospective crime mapping [e.g. Bowers et al. (Br J Criminol 44:641-658, 2004)], and compare its performance against grid-based alternatives. We also examine how measures of predictive accuracy can be translated to the network context, and show how differences in performance between the two cases can be quantified and tested. Results Findings demonstrate that the calibrated network-based model substantially outperforms a grid-based alternative in terms of predictive accuracy, with, for example, approximately 20 % more crime identified at a coverage level of 5 %. The improvement in accuracy is highly statistically significant at all coverage levels tested (from 1 to 10 %). Conclusions This study suggests that, for property crime at least, network-based methods of crime forecasting are likely to outperform grid-based alternatives, and hence should be & Toby Davies
BackgroundEvery day, around 400 million tweets are sent worldwide, which has become a rich source for detecting, monitoring and analysing news stories and special (disaster) events. Existing research within this field follows key words attributed to an event, monitoring temporal changes in word usage. However, this method requires prior knowledge of the event in order to know which words to follow, and does not guarantee that the words chosen will be the most appropriate to monitor.MethodsThis paper suggests an alternative methodology for event detection using space-time scan statistics (STSS). This technique looks for clusters within the dataset across both space and time, regardless of tweet content. It is expected that clusters of tweets will emerge during spatio-temporally relevant events, as people will tweet more than expected in order to describe the event and spread information. The special event used as a case study is the 2013 London helicopter crash.Results and ConclusionA spatio-temporally significant cluster is found relating to the London helicopter crash. Although the cluster only remains significant for a relatively short time, it is rich in information, such as important key words and photographs. The method also detects other special events such as football matches, as well as train and flight delays from Twitter data. These findings demonstrate that STSS is an effective approach to analysing Twitter data for event detection.
Landslides are one of the most destructive natural hazards; they can drastically alter landscape morphology, destroy man-made structures, and endanger people's life. Landslide susceptibility maps (LSMs), which show the spatial likelihood of landslide occurrence, are crucial for environmental management, urban planning, and minimizing economic losses. To date, the majority of research into data mining LSM uses small-scale case studies focusing on a single type of landslide. This paper presents a data mining approach to producing LSM for a large, heterogeneous region that is susceptible to multiple types of landslides. Using a case study of Piedmont, Italy, a Random Forest algorithm is applied to produce both susceptibility maps and classification maps. These maps are combined to give a highly accurate (over 85% classification accuracy) LSM which contains a large amount of information and is easy to interpret. This novel method of mapping landslide susceptibility demonstrates the efficacy of Random Forest to produce highly accurate susceptibility maps for a large heterogeneous region without the need for multiple susceptibility assessments.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
334 Leonard St
Brooklyn, NY 11211
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.