Table 2. Summary of each machine learning model employed in our study, reference and reasoning for application. |
Machine Learning
Model |
Abbreviation |
Reference |
Reasoning |
K-Nearest-Neighbors
regressor |
KNeighbors |
(Fix and
Hodges, Jr, J. L., 1951) |
K-nearest
neighbor (KNN) models are particularly effective in regression tasks where
local data patterns are crucial, as they base predictions on the proximity
to neighboring data points. |
Ridge |
Ridge |
(Friedman
et al., 2010) |
Linear
regression based models are beneficial in regression tasks, particularly
when regularization is required to improve model generalization and mitigate
overfitting. |
Lasso |
Lasso |
|
Elastic-Net |
ElasticNet |
(Friedman
et al., 2010) |
Decision
tree |
DecisionTree |
(Breiman,
2017) |
Tree-based
models are highly effective in regression tasks, especially in scenarios
where complex, non-linear relationships or intricate interactions exist
between the input features and the target variable. |
Extremely
randomized trees |
ExtraTree |
(Geurts
et al., 2006) |
Random forest |
RandomForest |
(Breiman,
2001) |
Gradient
tree boosting |
GradientBoosting |
(Friedman,
2002) |
Adaptive
boosting |
AdaBoost |
(Freund and
Schapire, 1995) |
|
|