2014
DOI: 10.1590/1413-81232014194.01722013
|View full text |Cite
|
Sign up to set email alerts
|

Método de mineração de dados para identificação de câncer de mama baseado na seleção de variáveis

Abstract: Método de mineração de dados para identificação de câncer de mama baseado na seleção de variáveisA data mining method for breast cancer identification based on a selection of variables

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0
2

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 20 publications
0
2
0
2
Order By: Relevance
“…These image processing steps include morphological feature computing (Sahiner et al 2001, Chen et al 2003, Armato and Sensakovic 2004, Joo et al 2004, Way et al 2006, Cheng et al 2010, which is still difficult to solve (Arbelaez et al 2011), and image decomposition (Sun et al 2013, Yang et al 2013, followed by statistical summaries and presentations (Sorensen 2010, Gmez et al 2012, Sun et al 2013, Yang et al 2013 for the calculation of textural features (Tourassi 1999). In feature integration by classifier, the widely used techniques are based on the KNN (k-nearest neighbor) method (Sahan et al 2007, Holsbach et al 2014, LDA (linear discriminant analysis) (Perez et al 2013(Perez et al , 2014 and SVM (support vector machines) (Akay et al 2009, Wang et al 2009, Krishnan et al 2010. The extraction of meaningful features is highly dependent on the quality of each intermediate result in the image processing steps (Tourassi 1999, Sahiner et al 2001, Joo et al 2004, Armato and Sensakovic 2004, Cheng et al 2010, which often requires recursive trial and error to obtain satisfactory results.…”
Section: Introductionmentioning
confidence: 99%
“…These image processing steps include morphological feature computing (Sahiner et al 2001, Chen et al 2003, Armato and Sensakovic 2004, Joo et al 2004, Way et al 2006, Cheng et al 2010, which is still difficult to solve (Arbelaez et al 2011), and image decomposition (Sun et al 2013, Yang et al 2013, followed by statistical summaries and presentations (Sorensen 2010, Gmez et al 2012, Sun et al 2013, Yang et al 2013 for the calculation of textural features (Tourassi 1999). In feature integration by classifier, the widely used techniques are based on the KNN (k-nearest neighbor) method (Sahan et al 2007, Holsbach et al 2014, LDA (linear discriminant analysis) (Perez et al 2013(Perez et al , 2014 and SVM (support vector machines) (Akay et al 2009, Wang et al 2009, Krishnan et al 2010. The extraction of meaningful features is highly dependent on the quality of each intermediate result in the image processing steps (Tourassi 1999, Sahiner et al 2001, Joo et al 2004, Armato and Sensakovic 2004, Cheng et al 2010, which often requires recursive trial and error to obtain satisfactory results.…”
Section: Introductionmentioning
confidence: 99%
“…A wide variety of machine learning classifiers have been developed for early diagnosis of breast cancer 16 17 18 . The widely used techniques are based on support vector machines (SVM) 18 19 20 , k-nearest neighbor (KNN) method 21 22 and linear discriminant analysis (LDA) 23 24 . However, the discriminative power of these methods is limited due to the computational costs of identifying definitive features for subset characterization and optimization.…”
mentioning
confidence: 99%
“…A primeira consistiu na validac ¸ão cruzada pelo técnica k -fold , que utilizou toda a base de dados e consistiu em realizar 'k' partic ¸ões (1 para teste e k-1 para treino), alternando os dados de treino e teste por 'k' vezes. Fez-se uso do k = 10 em todos os 6 modelos de classificac ¸ão [13]. Já na segunda configurac ¸ão, a base de dados foi dividida entre treino e teste, na seguinte proporc ¸ão: 20% para teste e 80% para o treinamento, pelo comando 'split', com uso do parâmetro 'stratify' e 'random state', sendo este último igual a 300 (escolha aleatória), para garantir que, cada vez que o código fosse inicializado, a divisão fosse a mesma.…”
Section: B Validac ¸ãO Cruzada E Divisão Treinamento/testeunclassified