2006
DOI: 10.1007/s10994-006-8713-9
|View full text |Cite
|
Sign up to set email alerts
|

First order random forests: Learning relational classifiers with complex aggregates

Abstract: In relational learning, predictions for an individual are based not only on its own properties but also on the properties of a set of related individuals. Relational classifiers differ with respect to how they handle these sets: some use properties of the set as a whole (using aggregation), some refer to properties of specific individuals of the set, however, most classifiers do not combine both. This imposes an undesirable bias on these learners. This article describes a learning approach that avoids this bia… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
56
0

Year Published

2006
2006
2022
2022

Publication Types

Select...
7

Relationship

3
4

Authors

Journals

citations
Cited by 50 publications
(56 citation statements)
references
References 30 publications
0
56
0
Order By: Relevance
“…Van Assche et al [21] have made the first implementation of combined aggregates and selections in an ILP system. They have extended the refinement operator of the relational decision tree learner Tilde [3] to include so-called complex aggregates: literals of the form F (V, Q, R), where F is an aggregate function (e.g., count), V is an aggregate variable occurring in the aggregate query Q, and R is the result of applying F to the set of all answer substitutions for V that Q results in (we will call this set the result set of Q).…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Van Assche et al [21] have made the first implementation of combined aggregates and selections in an ILP system. They have extended the refinement operator of the relational decision tree learner Tilde [3] to include so-called complex aggregates: literals of the form F (V, Q, R), where F is an aggregate function (e.g., count), V is an aggregate variable occurring in the aggregate query Q, and R is the result of applying F to the set of all answer substitutions for V that Q results in (we will call this set the result set of Q).…”
Section: Related Workmentioning
confidence: 99%
“…A complex aggregate can be constructed by iterative refinement of Q, starting with a general query (e.g., the number of atoms of a molecule) and ending with a very specific one (e.g., the number of carbon atoms bound with an aromatic bond type to an atom with charge larger than 0.06). The feature explosion resulting from combining aggregate functions with selection conditions was handled by upgrading Tilde to a random forest [21] and taking advantage of the feature sampling applied at each node of the trees. Charnay et al [6] have recently proposed an alternative solution, by introducing a hill-climbing approach to build complex aggregates incrementally.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The formatting of the feature value emphasizes the hierarchical, combinatorial structure of these count of count values (which can also be nested deeper into count of count of count values, etc.). Note that count of count features in our sense are quite different from nested aggregates in the sense of [1]: unlike the latter, our counts of counts do not aggregate a multiset of values into a single number at each level of nesting.…”
Section: Introductionmentioning
confidence: 97%
“…Perhaps the count of actors with a proportion of ≥ 50% award winning movies among the movies they participated in is the more relevant feature (here evaluating to 0 and 2, respectively, assuming that the movie to be classified is not considered in the count). Nested aggregates like this have been used in conjunction with first-order decision trees [1]. This very simple example illustrates three interesting aspects of relational features.…”
Section: Introductionmentioning
confidence: 99%