Decision Tree Model for Classifying University Students Eligible for UKT Waivers

Abstract views: 6 , PDF downloads: 2

Gde Yoga Agastyar Priatdana

Abstract

This paper develops a Decision Tree-based classification model to determine student eligibility for UKT (Single Tuition Fee) waivers using socio-economic factors such as parental income, household type, parental occupation, number of dependents, and vehicle ownership. The goal is to automate the identification of students qualifying for financial aid, improving efficiency and fairness in resource allocation. The model was trained on a dataset containing both categorical and numerical features, with the target variable being binary: "Eligible" (1) or "Not Eligible" (0). The model achieved an overall accuracy of 93.33%, with strong performance for the "Eligible" class, reflected by excellent precision, recall, and F1-score. However, the model performed poorly on the "Not Eligible" class, with low recall and F1-score, highlighting the issue of class imbalance. To address this, techniques like resampling and class weighting are recommended to improve classification of the minority class. Exploring alternative models like Random Forest or XGBoost could also provide more balanced results. This  underscores the importance of addressing class imbalance and using evaluation metrics beyond accuracy when developing classification models for imbalanced datasets.

Downloads

Download data is not yet available.
How to Cite
Agastyar Priatdana, G. (2025). Decision Tree Model for Classifying University Students Eligible for UKT Waivers. Jurnal Sistem Informasi Dan Komputer Terapan Indonesia (JSIKTI), 7(1), 285-294. https://doi.org/10.33173/jsikti.245

References

[1] L. Rokach and O. Maimon, "Decision Trees," in Data Mining and Knowledge Discovery Handbook, 2nd ed., Springer, 2015, pp. 165–192.
[2] W.-Y. Loh, "Regression Trees and Forests for Classification," Annals of Data Science, vol. 6, no. 1, pp. 1–26, Mar. 2019, doi: 10.1007/s40745-018-0161-9.
[3] L. Breiman, Classification and Regression Trees, New York, NY, USA: Routledge, 2017.
[4] I. H. Witten, E. Frank, M. A. Hall, and C. J. Pal, Data Mining: Practical Machine Learning Tools and Techniques, 5th ed., Burlington, MA, USA: Morgan Kaufmann, 2018.
[5] J. Brownlee, Imbalanced Classification with Python: Better Metrics, Balance Skewed Classes, and Improve Model Performance, 1st ed., Melbourne, Australia: Machine Learning Mastery, 2018.
[6] R. E. Schapire, "Explaining AdaBoost," in Empirical Inference, Springer, 2015, pp. 37–52, doi: 10.1007/978-3-662-45620-3_2.
[7] A. Liaw and M. Wiener, "Classification and Regression by RandomForest," R News, vol. 2, no. 3, pp. 18–22, Dec. 2018.
[8] C. Zhang, Y. Wang, and J. Chen, "An Explainable Machine Learning Framework for Imbalanced Data Classification," IEEE Access, vol. 7, pp. 165841–165851, Nov. 2019, doi: 10.1109/ACCESS.2019.2952862.
[9] H. Liu, J. Wu, and F. Zhang, "Feature Selection and Classification Using Decision Tree Optimized with Ensemble Methods," Journal of Systems and Software, vol. 134, pp. 138–148, Apr. 2018, doi: 10.1016/j.jss.2017.10.027.
[10] F. Pedregosa et al., "Scikit-learn: Machine Learning in Python," Journal of Machine Learning , vol. 12, pp. 2825–2830, Oct. 2019.