Proceedings of the International Workshop on Semantic Big Data 2016
DOI: 10.1145/2928294.2928305
|View full text |Cite
|
Sign up to set email alerts
|

Scalable algorithms for scholarly figure mining and semantics

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(29 citation statements)
references
References 12 publications
0
28
1
Order By: Relevance
“…While we did not directly compare our text role classifier with other approaches, we obtain a F1-score of 98%. This result is higher than the 92% F1-score reported by Choudhury et al [CWG16] using different features and a smaller data set (165 charts, 4,363 boxes). Our mark type classification approach exhibits superior performance to ReVision's classifier (F1-scores 94% vs. 80%) on their corpus.…”
Section: Discussioncontrasting
confidence: 68%
See 1 more Smart Citation
“…While we did not directly compare our text role classifier with other approaches, we obtain a F1-score of 98%. This result is higher than the 92% F1-score reported by Choudhury et al [CWG16] using different features and a smaller data set (165 charts, 4,363 boxes). Our mark type classification approach exhibits superior performance to ReVision's classifier (F1-scores 94% vs. 80%) on their corpus.…”
Section: Discussioncontrasting
confidence: 68%
“…In DiagramFlyer, Chen et al [CCA15] use feature vectors based only on text bounding boxes. Choudhury et al [CWG16] use text bounding boxes and also text content. We follow a similar approach, but do not incorporate text content, as it may propagate OCR errors.…”
Section: Miles/gallon Vs Horsepowermentioning
confidence: 99%
“…Thus, we discuss feature engineering and feature learning approaches used to extract visualization features. [11], [47], [54], [66] element positions or regions [44], [46], [53], [54], [67] element styles [46] parameters [1], [2] communicative signals [68] design rules [30], [36] statistical models [3], [66], [69], [70] statistics [31], [32], [64], [71], [72] one-hot vector [73] Feature Learning convolutional neural network [12], [13], [40], [44], [50], [53], [69], [74]- [77], [77]- [87] autoencoder [6], [77] autoencoder [9], [88] embedding models [3], [73] autoencoder [89]…”
Section: Representationmentioning
confidence: 99%
“…For our discussion, we classify existing approaches according to the feature space, including graphics, program, text, and underlying data (Table 1). It should be noted that some papers use multiple features for different tasks [44], [53] or use hybrids by feature fusion for improving performances [54], [66], [69]. In the following text, we describe each category in detail.…”
Section: Feature Engineering and Feature Learningmentioning
confidence: 99%
“…It usually combines text detection, optical character recognition (OCR), keypoint analysis, and other advanced computer vision methods [ 41 ]. Neural networks have shown surprising performance at this task [ 42 , 43 ]. Depending on the type of graph, specialized data extraction techniques exist (e.g., bar charts or pie charts use different methods; see [ 44 ]).…”
Section: Literature Reviewmentioning
confidence: 99%