2014
DOI: 10.1089/big.2014.0010
|View full text |Cite
|
Sign up to set email alerts
|

Bigger is Better, but at What Cost?Estimating the Economic Value of Incremental Data Assets

Abstract: Many firms depend on third-party vendors to supply data for commercial predictive modeling applications. An issue that has received very little attention in the prior research literature is the estimation of a fair price for purchased data. In this work we present a methodology for estimating the economic value of adding incremental data to predictive modeling applications and present two cases studies. The methodology starts with estimating the effect that incremental data has on model performance in terms of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 19 publications
0
5
0
Order By: Relevance
“…Then, f (Λ) ≥ 0 measures the extra surplus the firm can generate by using the data contained in Λ compared to a situation in which no data are available (i.e., f (0) = 0). To an extent, this approach is consistent with that of Dalessandro et al (2014) and the value function f (•) can be interpreted as the monetary evaluation of the dataset from the perspective of the data buyer.…”
Section: The Data Brokersmentioning
confidence: 70%
See 1 more Smart Citation
“…Then, f (Λ) ≥ 0 measures the extra surplus the firm can generate by using the data contained in Λ compared to a situation in which no data are available (i.e., f (0) = 0). To an extent, this approach is consistent with that of Dalessandro et al (2014) and the value function f (•) can be interpreted as the monetary evaluation of the dataset from the perspective of the data buyer.…”
Section: The Data Brokersmentioning
confidence: 70%
“…In an alternative example, this data structure also arises when the marginal contribution to the value of existing data is sufficiently low (Varian, 2018). For instance, Dalessandro et al (2014) show that the combination of an existing dataset with data from a third party may lead to near-zero marginal contribution, thereby rendering the merger between datasets almost useless. This is consistent with some observations made in Lambrecht and Tucker (2017).…”
Section: The Data Brokersmentioning
confidence: 99%
“…The data structure is extreme sub-additive when the value of the merged dataset is lower than the value of an individual dataset. For instance, Dalessandro et al (2014) suggest that, in some circumstances, adding additional data may be detrimental, and better predictions can be made with fewer data points. This is consistent with the seminal findings of Radner and Stiglitz (1984) who show theoretically that information can have a negative marginal net value.…”
Section: The Modelmentioning
confidence: 99%
“…Indeed, the number of examples tends to be the single most important factor in successful applications of machine learning, explaining much of the excitement about big data among practitioners. 10 It also needs a diverse set of examples. Stated simply, the computer will not be able to learn from examples to which it has not been exposed.…”
Section: Big Datamentioning
confidence: 99%