Ensemble Machine Learning Model Improves Prediction Accuracy for Academic Performance: A Comparative Study

of Default ML VS Boosting Algorithm
Like

Share this post

Choose a social network to share with, or copy the URL to share elsewhere

This is a representation of how your post may appear on social media. The actual post will vary between social networks

Overfitting is a common issue in machine learning models, where the model only predicts training data, but can be resolved using ensemble techniques like bagging or boosting. Bagging involves predicting accuracy through majority voting of multiple models running in parallel, while boosting uses a weak sequential model and passes the previously overfitted samples to another model to generate a more robust prediction. This research tries to fill the research gap between the selection of the best model for classification problem after comparing accuracy score of default and boosting algorithm. The multiple linear regression summary statistics indicate the dependent variables were significant, so prediction accuracy needs to be further tested using an ensemble machine learning model. The logistic regression model scored the least in both default (72%) and cross-validation (84%) accuracy than the random forest