Microblogging is a new form of communication in which users can describe their current status in short posts distributed by instant messages, mobile phones, email or the Web. Twitter, a popular microblogging tool has seen a lot of growth since it launched in October, 2006. In this paper, we present our observations of the microblogging phenomena by studying the topological and geographical properties of Twitter's social network. We find that people use microblogging to talk about their daily activities and to seek or share information. Finally, we analyze the user intentions associated at a community level and show how users with similar intentions connect with each other.
Social media systems such as weblogs, photo-and linksharing sites, wikis and on-line forums are currently thought to produce up to one third of new Web content. One thing that sets these "Web 2.0" sites apart from traditional Web pages and resources is that they are intertwined with other forms of networked data. Their standard hyperlinks are enriched by social networks, comments, trackbacks, advertisements, tags, RDF data and metadata. We describe recent work on building systems that use models of the Blogosphere to recognize spam blogs, find opinions on topics, identify communities of interest, derive trust relationships, and detect influential bloggers.
Abstract-We discuss the challenges involved in adapting the OntoSem natural language processing system to the Web. One set of tasks involves processing Web documents, translating their computed meaning representations from the OntoSem's native KR language into the Semantic Web language OWL, and publishing the results as Web pages and RSS feeds. Another set of tasks works in reverse -querying the Web for facts needed by OntoSem, translating them from OWL into OntoSem's native KR language and importing the results. A central problem underlying both sets of tasks is that of translating knowledge between OntoSem's KR language and ontologies and those of the Semantic Web. OntoSem2OWL has been developed as a translation system to support these translations. We describe SemNews, an implemented prototype application that demonstrates the process. It monitors RSS feeds of news stories, applies OntoSem to understand the text, and exports the computed facts back to the Web in OWL.
We describe our research on automatically generating rich semantic annotations of text and making it available on the Semantic Web. In particular, we discuss the challenges involved in adapting the OntoSem natural language processing system for this purpose. OntoSem, an implementation of the theory of ontological semantics under continuous development for over fifteen years, uses a specially constructed NLP-oriented ontology and an ontologicalsemantic lexicon to translate English text into a custom ontology-motivated knowledge representation language, the language of text meaning representations (TMRs). OntoSem concentrates on a variety of ambiguity resolution tasks as well as processing unexpected input and reference. To adapt OntoSem's representation to the Semantic Web, we developed a translation system, OntoSem2OWL, between the TMR language into the Semantic Web language OWL. We next used OntoSem and OntoSem2OWL to support SemNews, an experimental web service that monitors RSS news sources, processes the summaries of the news stories and publishes a structured representation of the meaning of the text in the news story. 2 The alternative on the supply side is, then, automatic annotation. Within the current state of the art, automatically produced annotations are roughly at the level attainable by the latest information extraction techniques -a reasonably good level of capturing named entities with a somewhat less successful categorization of such entities (e.g., decidingwhether Jordan is used as the first name of an individual or a reference to the Hashemite kingdom). Extracting more advanced types of semantic information, for example, types of events (to say nothing about determining semantic arguments, "case roles" in AI terminology), is not quite within the current information extraction capabilities, though work in this direction is ongoing. Indeed, semantic annotation is at the moment an active subfield of computational linguistics, where annotated corpora are intended for use by machine learning approaches to building natural language processing capabilities.On the demand side of the Semantic Web, a core capability is improving the precision of the Web search which will be facilitated by detailed semantic annotations that are unambiguous and sufficiently detailed to support the search engine in making fine-grained distinctions in calculating scores of documents. Another core capability is to transcend the level of document retrieval and instead return as answers to user queries specially generated pragmatically and stylistically appropriate responses. To attain this capability, intelligent agents must rely on very detailed semantic annotations of texts. We believe that such annotations will be, for all intents and purposes, complete text meaning representations, not just sets of semantic or pragmatic markers (and certainly not templates filled with uninterpreted snippets of the input text that are generated by the current information extraction methods).To attain such goals, Semantic Web agents must be ...
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.