Search citation statements
Paper Sections
Citation Types
Year Published
Publication Types
Relationship
Authors
Journals
With its rapid growth and increasing adoption, big data is producing a substantial impact in society. Its usage is opening both opportunities such as new business models and economic gains and risks such as privacy violations and discrimination. Europe is in need of a comprehensive strategy to optimise the use of data for a societal benefit and increase the innovation and competitiveness of its productive activities. In this paper, we contribute to the definition of this strategy with a research roadmap to capture the economic, social and ethical, legal and political benefits associated with the use of big data in Europe. The present roadmap considers the positive and negative externalities associated with big data, maps research and innovation topics in the areas of data management, processing, analytics, protection, visualisation, as well as non-technical topics, to the externalities they can tackle, and provides a time frame to address these topics in order to deliver social impact, skills development and standardisation. Finally, it also identifies what sectors will be most benefited by each of the research efforts. The goal of the roadmap is to guide European research efforts to develop a socially responsible big data economy, and to allow stakeholders to identify and meet big data challenges and proceed with a shared understanding of the societal impact, positive and negative externalities and concrete problems worth investigating in future programmes. IntroductionThe volume of data is growing exponentially, and is expected to reach the tens of zettabytes in 2020, of which a third is expected to be valuable if analysed, and about 40% will require protection [1]. The acquisition, analysis, curation, storage and usage of such big data may result in effects experienced by third parties that had no direct involvement in the activity itself. These externalities-positive if the action causes a positive effect or benefit to the third party, negative if it causes cost or harm-arise from decisions, activities or products by stakeholders such as industry, researchers and policy-makers.The present document contributes to the formulation of a strategy to define research and innovation efforts necessary for the realisation of a European big data economy by capturing and addressing the positive and negative societal externalities associated with the use of big data. It complements the technical challenges already identified [2] by taking a special focus on societal impacts, skills development and standardisation, and has been developed in parallel with a policy roadmap in the context of a multi-disciplinary study of the societal impacts of big data in seven European sectors aimed to define a roadmap and create a community that address and optimise these impacts [3].The term big data has received numerous definitions [4,5]. To develop the roadmap, we considered as a working definition that big data is that which uses big volume, big velocity, big variety data assets to extract value (insight and knowledge), and furthermor...
With its rapid growth and increasing adoption, big data is producing a substantial impact in society. Its usage is opening both opportunities such as new business models and economic gains and risks such as privacy violations and discrimination. Europe is in need of a comprehensive strategy to optimise the use of data for a societal benefit and increase the innovation and competitiveness of its productive activities. In this paper, we contribute to the definition of this strategy with a research roadmap to capture the economic, social and ethical, legal and political benefits associated with the use of big data in Europe. The present roadmap considers the positive and negative externalities associated with big data, maps research and innovation topics in the areas of data management, processing, analytics, protection, visualisation, as well as non-technical topics, to the externalities they can tackle, and provides a time frame to address these topics in order to deliver social impact, skills development and standardisation. Finally, it also identifies what sectors will be most benefited by each of the research efforts. The goal of the roadmap is to guide European research efforts to develop a socially responsible big data economy, and to allow stakeholders to identify and meet big data challenges and proceed with a shared understanding of the societal impact, positive and negative externalities and concrete problems worth investigating in future programmes. IntroductionThe volume of data is growing exponentially, and is expected to reach the tens of zettabytes in 2020, of which a third is expected to be valuable if analysed, and about 40% will require protection [1]. The acquisition, analysis, curation, storage and usage of such big data may result in effects experienced by third parties that had no direct involvement in the activity itself. These externalities-positive if the action causes a positive effect or benefit to the third party, negative if it causes cost or harm-arise from decisions, activities or products by stakeholders such as industry, researchers and policy-makers.The present document contributes to the formulation of a strategy to define research and innovation efforts necessary for the realisation of a European big data economy by capturing and addressing the positive and negative societal externalities associated with the use of big data. It complements the technical challenges already identified [2] by taking a special focus on societal impacts, skills development and standardisation, and has been developed in parallel with a policy roadmap in the context of a multi-disciplinary study of the societal impacts of big data in seven European sectors aimed to define a roadmap and create a community that address and optimise these impacts [3].The term big data has received numerous definitions [4,5]. To develop the roadmap, we considered as a working definition that big data is that which uses big volume, big velocity, big variety data assets to extract value (insight and knowledge), and furthermor...
Current Internet of Things (IoT) scenarios have to deal with many challenges especially when a large amount of heterogeneous data sources are integrated, that is, data curation. In this respect, the use of poor‐quality data (i.e., data with problems) can produce terrible consequence from incorrect decision‐making to damaging the performance in the operations. Therefore, using data with an acceptable level of usability has become essential to achieve success. In this article, we propose an IoT‐big data pipeline architecture that enables data acquisition and data curation in any IoT context. We have customized the pipeline by including the DMN4DQ approach to enable us the measuring and evaluating data quality in the data produced by IoT sensors. Further, we have chosen a real dataset from sensors in an agricultural IoT context and we have defined a decision model to enable us the automatic measuring and assessing of the data quality with regard to the usability of the data in the context.
Classical Datamining methods are facing various challenges in the era of Big Data. Between the need of fast knowledge extraction and the high flows of data acquired in small slots of time, these methods became shifted. The variability and the veracity of the Big Data perplex the Machine Learning process. The high volume of Big Data yields to a congested learning because the classic methods are designed for small sets of features. Deep Learning has recently emerged in the aim of handling voluminous data. The concept of the Deep induces the conversion of the features into a new abstracted representation in order to optimize an objective. Although the Deep Learning methods are experimentally promising, their parameterization is exhaustive and empirical.To tackle these problems, we utilize the causality and the uncertainty of the Bayesian Network in order to propose a new Deep Bayesian Network architecture. We provide a new learning algorithm for this multi-layered Bayesian Network with latent variables. We evaluate the proposed architecture and learning algorithms over benchmark datasets. We used high-dimensional data in order to simulate the Big Data challenges, which are imposed by the volume and veracity aspects. We demonstrate the effectiveness of our contribution under these constraints.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.