2021
DOI: 10.1002/smr.2367
|View full text |Cite
|
Sign up to set email alerts
|

An empirical comparison of validation methods for software prediction models

Abstract: Model validation methods (e.g., k-fold cross-validation) use historical data to predict how well an estimation technique (e.g., random forest) performs on the current (or future) data. Studies in the contexts of software development effort estimation (SDEE) and software fault prediction (SFP) have used and investigated different model validation methods. However, no conclusive indications to suggest which model validation method has a major impact on the prediction accuracy and stability of estimation techniqu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 62 publications
0
4
0
Order By: Relevance
“…Examples of hyper parameters include the number of epochs, batch size, and learning rate. Adjusting these hyper parameters may significantly impact the performance of a machine learning model and finding optimal values for a given problem is typically a critical step in the model development procedure [45]. Several tests were conducted before the current study to determine the appropriate value range.…”
Section: Rnn-based Deep Learning (Rnnbdl) Approachmentioning
confidence: 99%
“…Examples of hyper parameters include the number of epochs, batch size, and learning rate. Adjusting these hyper parameters may significantly impact the performance of a machine learning model and finding optimal values for a given problem is typically a critical step in the model development procedure [45]. Several tests were conducted before the current study to determine the appropriate value range.…”
Section: Rnn-based Deep Learning (Rnnbdl) Approachmentioning
confidence: 99%
“…The selection of appropriate parameters is critical and a complicated aspect of network training due to constraints such as memory limitations, trade-offs are inherently present in parameter selection [3], [47]. Throughout the examination, several hyper-parameters to measure the influence of accuracy.…”
Section: Hyper Parametermentioning
confidence: 99%
“…Nested K fold cross validation is another way to tune parameters of an algorithms. Data is divided into k fold and one fold is reserved for test [3]. K-1 training folds used for validation.…”
Section: Cross Validationmentioning
confidence: 99%
“…Model validation is the process where the trained model is evaluated with a testing data set to foresee how good the performance of the estimation method [59]. The testing data set is a separate portion of the same data set from which the training set is derived.…”
Section: Model Validationmentioning
confidence: 99%