2022
DOI: 10.1163/23641177-bja10055
|View full text |Cite
|
Sign up to set email alerts
|

Random Forest Analysis of Factors Predicting Science Achievement Groups: Focusing on Science Activities and Learning in School

Abstract: This study explored science-related variables that have an impact on the prediction of science achievement groups by applying the educational data mining (EDM) method of the random forest analysis to extract factors associated with students categorized in three different achievement groups (high, moderate, and low) in the Korean data from the 2015 Programme for International Student Assessment (PISA). The 57 variables of science activities and learning in school collected from PISA questionnaires for students … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 29 publications
0
0
0
Order By: Relevance
“…In recent years, machine learning methods have gained traction for analyzing extensive educational assessment data, estimating the influence of individual factors, and constructing predictive models [40]. Among these techniques, the RF approach has been proposed to assess the significance of individual factors in predicting outcomes [18,19,21].…”
Section: Research Aimsmentioning
confidence: 99%
See 1 more Smart Citation
“…In recent years, machine learning methods have gained traction for analyzing extensive educational assessment data, estimating the influence of individual factors, and constructing predictive models [40]. Among these techniques, the RF approach has been proposed to assess the significance of individual factors in predicting outcomes [18,19,21].…”
Section: Research Aimsmentioning
confidence: 99%
“…Within the present study, 20% of the samples were designated for validation data; tree numbers were fixed to 500, 1000, and 2000 trees; each tree drew on 50% of the training data. Two variables (the square root of the total number of variables, i.e., √ 5 ≈ 2) were randomly selected for node splitting in each decision tree generated from bootstrapped datasets [40,53]. Therefore, n(Train) = 17,609; n(Test) = 4402.…”
Section: Random Forestmentioning
confidence: 99%