2015
DOI: 10.1186/s12916-015-0521-2
|View full text |Cite
|
Sign up to set email alerts
|

Naive Bayes classifiers for verbal autopsies: comparison to physician-based classification for 21,000 child and adult deaths

Abstract: BackgroundVerbal autopsies (VA) are increasingly used in low- and middle-income countries where most causes of death (COD) occur at home without medical attention, and home deaths differ substantially from hospital deaths. Hence, there is no plausible “standard” against which VAs for home deaths may be validated. Previous studies have shown contradictory performance of automated methods compared to physician-based classification of CODs. We sought to compare the performance of the classic naive Bayes classifie… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
65
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(65 citation statements)
references
References 25 publications
0
65
0
Order By: Relevance
“…tuberculosis , which likely led to the exclusion of individuals with disseminated TB and limited or no respiratory symptoms. To date, all comparisons of VA to the PHMRC dataset, including those conducted by the PHMRC team, have combined the ‘AIDS’ and ‘AIDS with TB’ categories, and have therefore not attempted to assess VA’s ability to detect HIV-associated TB [19,20,30,6064]. The PHMRC gold standard dataset nevertheless remains a valuable resource; we would suggest that any future validation exercises use the differentiated, ‘AIDS with TB’ and ‘AIDS’ categories, rather than the combined ‘AIDS’ category, for comparison to VA.…”
Section: Discussionmentioning
confidence: 99%
“…tuberculosis , which likely led to the exclusion of individuals with disseminated TB and limited or no respiratory symptoms. To date, all comparisons of VA to the PHMRC dataset, including those conducted by the PHMRC team, have combined the ‘AIDS’ and ‘AIDS with TB’ categories, and have therefore not attempted to assess VA’s ability to detect HIV-associated TB [19,20,30,6064]. The PHMRC gold standard dataset nevertheless remains a valuable resource; we would suggest that any future validation exercises use the differentiated, ‘AIDS with TB’ and ‘AIDS’ categories, rather than the combined ‘AIDS’ category, for comparison to VA.…”
Section: Discussionmentioning
confidence: 99%
“…The openVA R-package (Li et al, 2018) has made many of these algorithms publicly available. Generic classifiers like random forests (Breiman, 2001), naive Bayes classifiers (Minsky, 1961) and support vector machines (Cortes and Vapnik, 1995) have also been used Miasnikof et al, 2015;Koopman et al, 2015). Estimated COD labels for each VA record in a nationally representative VA database is aggregated to obtain national cause specific mortality fractions (CSMF) -the population-level class membership probabilities, that are often the main quantities of interest for epidemiologists, local governments, and global health organizations.…”
Section: Motivating Datasetmentioning
confidence: 99%
“…This extension allows us to complete our model to simultaneously estimate the latent correlation matrix and assign causes of death using VA data. Before we describe our model, it is worth noting that for many existing automated VA methods such as InSilicoVA (McCormick et al, 2016), InterVA (Byass et al, 2003), and the Naive Bayes Classifier (Miasnikof et al, 2015), the classification rule is closely related to the naive Bayes classifier under the assumption of (conditional) independence between symptoms, i.e.…”
Section: Full Posterior Sampling Stepsmentioning
confidence: 99%
“…The majority of the existing statistical or algorithmic methods to assign cause of death using VA surveys make the assumption that VA symptoms are independent from one another conditional on cause of death (Byass et al, 2003;James et al, 2011;Miasnikof et al, 2015;McCormick et al, 2016). This assumption simplifies computation and is efficient in settings with limited training data.…”
Section: Introductionmentioning
confidence: 99%