2016
DOI: 10.7287/peerj.preprints.1713
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Similarity to a single set

Abstract: Abstract. Identifying patterns and associations in data is fundamental to discovery in science. This work investigates a very simple instance of the problem, where each data point consists of a vector of binary attributes, and attributes are treated equally. For example, each data point may correspond to a person and the attributes may be their sex, whether they smoke cigarettes, whether they have been diagnosed with lung cancer, etc. Measuring similarity of attributes in the data is equivalent to measuring si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2017
2017
2017
2017

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 16 publications
(43 reference statements)
0
3
0
Order By: Relevance
“…The first is to divide n f and e p by F and P , respectively (see Figure A). This results in metrics for which the number of passed and failed tests can be scaled separately without affecting the ranking, called “general scalable” in the work of Naish . The second scaling method is to divide both n f and e p by the total number of test cases T = F + P (see Figure B); this is also general scalable.…”
Section: The Hyperbolic Metric Classmentioning
confidence: 99%
See 2 more Smart Citations
“…The first is to divide n f and e p by F and P , respectively (see Figure A). This results in metrics for which the number of passed and failed tests can be scaled separately without affecting the ranking, called “general scalable” in the work of Naish . The second scaling method is to divide both n f and e p by the total number of test cases T = F + P (see Figure B); this is also general scalable.…”
Section: The Hyperbolic Metric Classmentioning
confidence: 99%
“…The third scaling method that we used in our experiments is to divide n f by F and e p by E (number of times a statement is executed, ie, e f + e p ; see Figure C). This form of scaling produces “uniform scalable” metrics, for which the scaling of passed and failed tests must be the same to ensure the same ranking. The class still has the desired properties (a combination of single‐bug optimal, deterministic bug optimal, and Ochiai‐ or Kulczynski2‐like contours).…”
Section: The Hyperbolic Metric Classmentioning
confidence: 99%
See 1 more Smart Citation