Abstract:A survey of over 400 recent research articles suggests that computer scientists publish relatively few papers with experimentally validated r esults.The survey includes complete volumes of several refereed c omputer science journals, a conference, and 50 titles drawn at random from all articles published b y A CM in 1993. The journals Optical Engineering OE and Neural Computation NC were used for comparison.Of the papers in the random sample that would require experimental validation, 40 have none at all. In j… Show more
“…For this purpose, we took the papers published at ISWC since 2002 and manually classified them according to the scheme proposed by [5]. In addition, we had a closer look at papers containing descriptions of experimental work with respect to the data used and the claims made.…”
Section: Research Questions and Methodsmentioning
confidence: 99%
“…In order to be able to compare our findings to previous studies on the role of experimental work in computer science as a whole, we used the classification scheme proposed in [5] with the modifications described in [11], i.e. the merge of the two categories 'Empirical Work' and 'Hypothesis Testing'.…”
Section: Annotation Schemementioning
confidence: 99%
“…While the amount of such papers is lower in certain subareas of computer science, like software engineering (55%), still a significant amount of work in computer science is formulative and requires some evaluation. So far, the most detailed and systematic investigation of experimental research as a means for evaluating formulative research has been carried out by Tichy and others in 1995 [5]. Based on a sample of publications from major computer science journals the authors categorize papers into formal theory, design and modeling, as well as empirical work and others.…”
Section: Empirical Studies Of Experimental Research In Computer Sciencementioning
confidence: 99%
“…As described above, we take the relative number of pages describing experiments (RELPAGES) as a general proxy for the importance of the experimental part within a paper. Following [5], we thus make the assumption that the better the experiments in a paper are, the more the paper reports about the experiments.…”
Section: H5 Strong Experimental Work Increases the Impact Of A Papermentioning
confidence: 99%
“…In particular, we present descriptive statistics of the ISWC paper collection and results of investigating possible correlations with research impact. [5] refer to papers published in 1995, and [6] used papers from 2005, our study covers 11 consecutive years from 2002 -2012. This also explains the comparably high number of papers (500) included in our study.…”
Abstract. Experimentation is an important way to validate results of Semantic Web and Computer Science research in general. In this paper, we investigate the development and the current status of experimental work on the Semantic Web. Based on a corpus of 500 papers collected from the International Semantic Web Conferences (ISWC) over the past decade, we analyse the importance and the quality of experimental research conducted and compare it to general Computer Science. We observe that the amount and quality of experiments are steadily increasing over time. Unlike hypothesised, we cannot confirm a statistically significant correlation between a paper's citations and the amount of experimental work reported. Our analysis, however, shows that papers comparing themselves to other systems are more often cited than other papers.
“…For this purpose, we took the papers published at ISWC since 2002 and manually classified them according to the scheme proposed by [5]. In addition, we had a closer look at papers containing descriptions of experimental work with respect to the data used and the claims made.…”
Section: Research Questions and Methodsmentioning
confidence: 99%
“…In order to be able to compare our findings to previous studies on the role of experimental work in computer science as a whole, we used the classification scheme proposed in [5] with the modifications described in [11], i.e. the merge of the two categories 'Empirical Work' and 'Hypothesis Testing'.…”
Section: Annotation Schemementioning
confidence: 99%
“…While the amount of such papers is lower in certain subareas of computer science, like software engineering (55%), still a significant amount of work in computer science is formulative and requires some evaluation. So far, the most detailed and systematic investigation of experimental research as a means for evaluating formulative research has been carried out by Tichy and others in 1995 [5]. Based on a sample of publications from major computer science journals the authors categorize papers into formal theory, design and modeling, as well as empirical work and others.…”
Section: Empirical Studies Of Experimental Research In Computer Sciencementioning
confidence: 99%
“…As described above, we take the relative number of pages describing experiments (RELPAGES) as a general proxy for the importance of the experimental part within a paper. Following [5], we thus make the assumption that the better the experiments in a paper are, the more the paper reports about the experiments.…”
Section: H5 Strong Experimental Work Increases the Impact Of A Papermentioning
confidence: 99%
“…In particular, we present descriptive statistics of the ISWC paper collection and results of investigating possible correlations with research impact. [5] refer to papers published in 1995, and [6] used papers from 2005, our study covers 11 consecutive years from 2002 -2012. This also explains the comparably high number of papers (500) included in our study.…”
Abstract. Experimentation is an important way to validate results of Semantic Web and Computer Science research in general. In this paper, we investigate the development and the current status of experimental work on the Semantic Web. Based on a corpus of 500 papers collected from the International Semantic Web Conferences (ISWC) over the past decade, we analyse the importance and the quality of experimental research conducted and compare it to general Computer Science. We observe that the amount and quality of experiments are steadily increasing over time. Unlike hypothesised, we cannot confirm a statistically significant correlation between a paper's citations and the amount of experimental work reported. Our analysis, however, shows that papers comparing themselves to other systems are more often cited than other papers.
SUMMARYEmpirical systems research is facing a dilemma. Minor aspects of an experimental setup can have a significant impact on its associated performance measurements and potentially invalidate conclusions drawn from them. Examples of such influences, often called hidden factors, include binary link order, process environment size, compiler generated randomized symbol names, or group scheduler assignments. The growth in complexity and size of modern systems will further aggravate this dilemma, especially with the given time pressure of producing results. How can one trust any reported empirical analysis of a new idea or concept in computer science? DataMill is a community-based services-oriented open benchmarking infrastructure for rigorous performance evaluation. DataMill facilitates producing robust, reliable, and reproducible results. The infrastructure incorporates the latest results on hidden factors and automates the variation of these factors. DataMill is also of interest for research on performance evaluation. The infrastructure supports quantifying the effect of hidden factors, disseminating the research results beyond mere reporting. It provides a platform for investigating interactions and composition of hidden factors. This paper discusses experience earned through creating and using an open benchmarking infrastructure. Multiple research groups participate and have used DataMill. Furthermore, DataMill has been used for a performance competition at the International Conference on Runtime Verification (RV) 2014 and is currently hosting the RV 2015 competition. This paper includes a summary of our experience hosting the first RV competition.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.