2013 Joint IFSA World Congress and NAFIPS Annual Meeting (IFSA/NAFIPS) 2013
DOI: 10.1109/ifsa-nafips.2013.6608465
|View full text |Cite
|
Sign up to set email alerts
|

How deep data becomes big data

Abstract: We present some problems and solutions for situations when compound and semantically rich nature of data records, such as scientific articles, creates challenges typical for big data processing. Using a case study of named entity matching in SONCA system we show how big data problems emerge and how they are solved by bringing together methods from database management and computational intelligence.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 14 publications
0
5
0
Order By: Relevance
“…The above case study is an example of utilization of a well-known strategy of solving complex computational tasks, where one first heuristically partitions a problem onto more feasible pieces, then conducts detailed computations over such pieces, and finally aggregates local results [18], [19]. In our case, partitioning is performed by the first layer of comparators over the Cartesian product A × B of input and reference objects, while detailed calculations refer to comparators in the second layer.…”
Section: A Network Of Monolithic Objectsmentioning
confidence: 99%
See 2 more Smart Citations
“…The above case study is an example of utilization of a well-known strategy of solving complex computational tasks, where one first heuristically partitions a problem onto more feasible pieces, then conducts detailed computations over such pieces, and finally aggregates local results [18], [19]. In our case, partitioning is performed by the first layer of comparators over the Cartesian product A × B of input and reference objects, while detailed calculations refer to comparators in the second layer.…”
Section: A Network Of Monolithic Objectsmentioning
confidence: 99%
“…Our system is called SONCA (Search based on ON tologies and C ompound Analytics; see e.g. [17], [18]). In SONCA, each acquired document is stored in a parsed form in a data warehouse called SoncaDB, in order to make the best possible use of the available structural and semantic information while searching, filtering, grouping, etc.…”
Section: A Network Of Monolithic Objectsmentioning
confidence: 99%
See 1 more Smart Citation
“…The term Deep Data describes data that is not very big, but semantically very rich, i.e., it provides contextually rich information intended to provide a rich user experience. The term Deep Data was proposed by analogy with the term Deep Web, which is used to describe the vast amount of information that resides beneath the surface of web pages (Szczuka & Ślȩzak, 2013). The term Deep Data has also been used to describe an approach where all the information available in the data is fully exploited to gain knowledge (Belianinov et al, 2015).…”
Section: Problemmentioning
confidence: 99%
“…• 2199 -one sliding window per short time series • 6315 -two sliding windows per short time series • 18663 -five sliding windows per short time series Making a total of 27177 attributes [27] from the conditionaland inter-sliding windows constructed for both raw and virtual sensors. Elements of the automatic feature selection and reducing the number of attributes are in the study phase and still have not been introduced to the data processing mechanisms.…”
Section: Post Processingmentioning
confidence: 99%