Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence 2019
DOI: 10.24963/ijcai.2019/272
|View full text |Cite
|
Sign up to set email alerts
|

Inter-node Hellinger Distance based Decision Tree

Abstract: This paper introduces a new splitting criterion called Inter-node Hellinger Distance (iHD) and a weighted version of it (iHDw) for constructing decision trees. iHD measures the distance between the parent and each of the child nodes in a split using Hellinger distance. We prove that this ensures the mutual exclusiveness between the child nodes. The weight term in iHDw is concerned with the purity of individual child node considering the class imbalance problem. The combination of the distance and weight term i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 8 publications
(1 reference statement)
0
6
0
Order By: Relevance
“…Motivated by HDT, several extensions have been proposed to handle different tasks under class imbalance: data streams (Grzyb, Klikowski, & Woźniak, 2021;Lyon et al, 2014), multilabel classification (Daniels & Metaxas, 2017), and multiclass classification (Hoens, Qian, Chawla, & Zhou, 2012). Other works have used the Hellinger distance to propose their own methods that aim to outperform the HDT (Akash, Kadir, Ali, & Shoyaib, 2019;Su & Cao, 2019). Despite the popularity of HDT in other domains, it is not yet explored in weakly supervised learning.…”
Section: Imbalanced Learningmentioning
confidence: 99%
“…Motivated by HDT, several extensions have been proposed to handle different tasks under class imbalance: data streams (Grzyb, Klikowski, & Woźniak, 2021;Lyon et al, 2014), multilabel classification (Daniels & Metaxas, 2017), and multiclass classification (Hoens, Qian, Chawla, & Zhou, 2012). Other works have used the Hellinger distance to propose their own methods that aim to outperform the HDT (Akash, Kadir, Ali, & Shoyaib, 2019;Su & Cao, 2019). Despite the popularity of HDT in other domains, it is not yet explored in weakly supervised learning.…”
Section: Imbalanced Learningmentioning
confidence: 99%
“…Akash et al [13] proposed a skew insensitive splitting criterion called Inter-node Hellinger Distance (iHD) that uses the squared Hillinger distance (D 2 H ) to measure the heterogeneity between the class probability distributions of the parent and child nodes.…”
Section: Inter-node Hellinger Distancementioning
confidence: 99%
“…Breiman and Friedman [2] introduced Classification And Regression Trees (CART) which provides the Gini Index and towing criterion as impurity measurement. More recently, Akash et al [13] proposed a splitting criterion called Inter-node Hellinger Distance (iHD), which measures the distance between parent and children nodes using Hellinger Distance. The splitting criteria of the above described and our proposal algorithms are compared in Table I below.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In contrast, algorithm oriented approaches are the modifications of traditional algorithms such as DT and kNN. The modified DTs for imbalanced classification are Hellinger Distance DT (HDDT) [5], Class Confidence Proportion DT (CCPDT) [13] and Weighted Inter-node Hellinger Distance DT (iHDwDT) [1]. These DTs use different splitting criteria while selecting a feature in split point.…”
Section: Introductionmentioning
confidence: 99%