2008
DOI: 10.1109/tvcg.2008.178
|View full text |Cite
|
Sign up to set email alerts
|

Vispedia: Interactive Visual Exploration of Wikipedia Data via Search-Based Integration

Abstract: Abstract-Wikipedia is an example of the collaborative, semi-structured data sets emerging on the Web. These data sets have large, non-uniform schema that require costly data integration into structured tables before visualization can begin. We present Vispedia, a Web-based visualization system that reduces the cost of this data integration. Users can browse Wikipedia, select an interesting data table, then use a search interface to discover, integrate, and visualize additional columns of data drawn from multip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
16
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 33 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…Furthermore, many extraction problems are by nature ambiguous and require user input, yet mostlyautomatic systems like Sifter [10] offer the user little help when the extractors fail. Karma [4] and Mashmaker [7] can learn from positive but not negative examples. Mashmaker users must drop into a lower level pattern editor to make a pattern more selective.…”
Section: Programmer Leverages Structural Heuristicsmentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, many extraction problems are by nature ambiguous and require user input, yet mostlyautomatic systems like Sifter [10] offer the user little help when the extractors fail. Karma [4] and Mashmaker [7] can learn from positive but not negative examples. Mashmaker users must drop into a lower level pattern editor to make a pattern more selective.…”
Section: Programmer Leverages Structural Heuristicsmentioning
confidence: 99%
“…For instance, Vispedia [4] allows visualization of Wikipedia articles by leveraging the RDF predefined for each topic as part of the DBpedia project. d.mix [12] allows experts to define a library of patterns that end users employ.…”
Section: Programmer Leverages Predefined Webpage Semanticsmentioning
confidence: 99%
“…For example, Sifter [16] extracts search items on a web page using the HTML structure and scrapes subsequent web pages by examining hyperlinks (such as "Next page") and URL parameters. Vispedia [17] extracts Wikipedia infoboxes using the table structure and uses the hyperlinks in an infobox to retrieve related topics. There are also commercial web scrapers, such as Scraper [18], a Chrome plugin for scraping similar items in web pages, and ScraperWiki [19], a commercial product that specifically targets scraping Twitter and tabular data.…”
Section: Related Workmentioning
confidence: 99%
“…The first allows the traversal of an RDF graph while the latter enables the analysis of aggregated data that match a query. The Vispedia project [6] is an approach to interactively visualize Wikipedia infoboxes. It allows a user to select an infobox and define a keyword query, which the system evaluates on the semantic graph of Wikipedia to extract supplemental information.…”
Section: Related Workmentioning
confidence: 99%
“…As for data profiling in general, there is a wealth of approaches that can be used to grasp a given dataset, e.g., functional dependency discovery [6], and join path exploration [11]. The Bellman project [12] integrates a set of techniques to address poorly structured and dirty data.…”
Section: Related Workmentioning
confidence: 99%