Classifying School Scope Using Deep Neural Networks Based on Students' Surrounding Living Environments
DOI:
https://doi.org/10.62411/jcta.11739Keywords:
Classification, Deep Neural Networks, Educational Analytics, Educational Interventions, Overfitting Prevention, School ScopeAbstract
This research investigates school scope classification using Deep Neural Networks (DNN), focusing on students living environments and educational opportunities. By addressing the interplay of socioeconomic and educational factors, the study aims to develop an analytical framework for understanding how environmental contexts shape academic trajectories. The research provides a nuanced understanding of the importance of features in educational classification by developing DNN models based on Spearman's Rank Correlation Coefficient (SRCC). The methodology employs machine learning techniques, integrating data wrangling, exploratory analysis, and multiple DNN models with K-fold cross-validation. The study analyzes 677 student records from two schools. The research examined multiple model configurations. Results show that the 'All Data' model achieved 83.08% accuracy, the 'Top 5' model 81.54%, and the 'Non-Top 5' model 79.23%. The SRCC-based approach revealed that while top correlated features are important, additional variables significantly contribute to model performance. The study highlights the profound impact of family background, social environment, and educational contexts on school selection. Furthermore, it demonstrates DNN's capability to uncover intricate, non-linear relationships, offering actionable insights for policymakers to leverage machine learning's potential in developing targeted educational strategies.References
J. Huang, W. Li, and C. Liu, “The Influence of Family Environment on Secondary School Students’ Academic Performance,” Lecture Notes in Education Psychology and Public Media, vol. 12, no. 1, pp. 327–335, Oct. 2023, doi: 10.54254/2753-7048/12/20230837.
Muzamil Hussain AL Hussaini and Tariq Hussain, “Impact of Environmental Factors on Student’s Educational Background and Their Future Performance,” International Journal of Scientific Multidisciplinary Research, vol. 1, no. 6, pp. 607–616, Jul. 2023, doi: 10.55927/ijsmr.v1i6.5021.
W. Yiting, “Study on the influencing factors of family environment on the learning habits of primary school students in middle school,” Applied & Educational Psychology, vol. 4, no. 11, pp. 6–10, Nov. 2023, doi: 10.23977/appep.2023.041102.
A. Zubair, “Klasifikasi Ketertarikan Mahasiswa Pembelajaran Berbasiskan Video Business Intelligence Menggunakan Adaboost,” in Seminar Nasional Sistem Informasi (SENASIF), 2022, pp. 3050–3058.
I. H. Nisa, “The Role Of The School Environment On Student Learning Success, A Descriptive Study at MTS Dzunnuraini,” TARQIYATUNA: Jurnal Pendidikan Agama Islam dan Madrasah Ibtidaiyah, vol. 1, no. 1, pp. 24–29, Mar. 2022, doi: 10.36769/tarqiyatuna.v1i1.190.
N. Walia, M. Kumar, N. Nayar, and G. Mehta, “Student’s Academic Performance Prediction in Academic using Data Mining Techniques,” Apr. 01, 2020, Social Science Research Network, Rochester, NY: 3565874. doi: 10.2139/ssrn.3565874.
F. Mustofa, A. N. Safriandono, A. R. Muslikh, and D. R. I. M. Setiadi, “Dataset and Feature Analysis for Diabetes Mellitus Classi-fication using Random Forest,” Journal of Computing Theories and Applications, vol. 1, no. 1, pp. 41–48, Jan. 2023, doi: 10.33633/jcta.v1i1.9190.
S. Mittal, “A survey on modeling and improving reliability of DNN algorithms and accelerators,” Journal of Systems Architecture, vol. 104, p. 101689, Mar. 2020, doi: 10.1016/j.sysarc.2019.101689.
C. Schröder and A. Niekler, “A Survey of Active Learning for Text Classification using Deep Neural Networks,” Aug. 17, 2020, arXiv. doi: 10.48550/arXiv.2008.07267.
B. Fu, Q. Shang, T. Sun, and S. Jia, “A Distracted Driving Detection Model Based On Driving Performance,” IEEE Access, vol. 11, pp. 26624–26636, 2023, doi: 10.1109/ACCESS.2023.3257238.
C. Janiesch, P. Zschech, and K. Heinrich, “Machine learning and deep learning,” Electronic Markets, vol. 31, no. 3, pp. 685–695, Sep. 2021, doi: 10.1007/s12525-021-00475-2.
H. Y. Song and S. Park, “An Analysis of Correlation between Personality and Visiting Place using Spearman’s Rank Correlation Coefficient,” KSII Transactions on Internet and Information Systems, vol. 14, no. 5, May 2020, doi: 10.3837/tiis.2020.05.005.
N. Khatun, “Applications of Normality Test in Statistical Analysis,” Open Journal of Statistics, vol. 11, no. 01, pp. 113–122, 2021, doi: 10.4236/ojs.2021.111006.
S. Korkmaz and Y. Demir, “Investigation of Some Univariate Normality Tests In Terms of Type-I Errors and Test Power,” Journal of Scientific Reports-A, no. 052, pp. 376–395, Mar. 2023, doi: 10.59313/jsr-a.1222979.
O. Chamorro-Atalaya et al., “K-Fold Cross-Validation through Identification of the Opinion Classification Algorithm for the Sat-isfaction of University Students,” International Journal of Online and Biomedical Engineering (iJOE), vol. 19, no. 11, Aug. 2023, doi: 10.3991/ijoe.v19i11.39887.
P. Cortez and A. M. G. Silva, “Using data mining to predict secondary school student performance,” Apr. 2008.
D. Marcondes, A. Simonis, and J. Barrera, “Feature Selection based on the Local Lift Dependence Scale,” Entropy, vol. 20, no. 2, p. 97, Jan. 2018, doi: 10.3390/e20020097.
A. B. F. Mansur and N. Yusof, “The Latent of Student Learning Analytic with K-mean Clustering for Student Behaviour Classifi-cation,” Journal of Information Systems Engineering and Business Intelligence, vol. 4, no. 2, p. 156, Oct. 2018, doi: 10.20473/jisebi.4.2.156-161.
R. Yacouby and D. Axman, “Probabilistic Extension of Precision, Recall, and F1 Score for More Thorough Evaluation of Classi-fication Models,” in Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems, Stroudsburg, PA, USA: Association for Computational Linguistics, 2020, pp. 79–91. doi: 10.18653/v1/2020.eval4nlp-1.9.
O. K. Oyedotun, S. N. Tackie, E. O. Olaniyi, and A. Khashman, “Data Mining of Students’ Performance: Turkish Students as a Case Study,” International Journal of Intelligent Systems and Applications, vol. 7, no. 9, pp. 20–27, Sep. 2015, doi: 10.5815/ijisa.2015.09.03.
M. Karagiannopoulos, D. Anyfantis, S. Kotsiantis, and P. Pintelas, “A Wrapper for Reweighting Training Instances for Handling Imbalanced Data Sets,” in Artificial Intelligence and Innovations 2007: from Theory to Applications, vol. 247, C. Boukis, A. Pnevmatikakis, and L. Polymenakos, Eds., Boston, MA: Springer US, 2007, pp. 29–36. doi: 10.1007/978-0-387-74161-1_4.
N. E. Helwig, “Adding bias to reduce variance in psychological results: A tutorial on penalized regression,” The Quantitative Methods for Psychology, vol. 13, no. 1, pp. 1–19, Jan. 2017, doi: 10.20982/tqmp.13.1.p001.
The pandas development team, “pandas-dev/pandas: Pandas,” Jun. 2023, Zenodo. doi: 10.5281/zenodo.8092754.
O. Grisel et al., “scikit-learn/scikit-learn: Scikit-learn 1.5.0,” Zenodo. Accessed: Jul. 13, 2024. [Online]. Available: https://zenodo.org/records/11237090
R. Gommers et al., “scipy/scipy: SciPy 1.11.4,” Zenodo. Accessed: Jul. 13, 2024. [Online]. Available: https://zenodo.org/records/10155614
The Matplotlib Development Team, “Matplotlib: Visualization with Python,” Zenodo. Accessed: Jul. 13, 2024. [Online]. Available: https://zenodo.org/records/11201097
P. Cortez, “Student Performance,” UC Irvine Machine Learning Repository. Accessed: Jul. 13, 2024. [Online]. Available: https://archive.ics.uci.edu/dataset/320
W. N. Wassouf, R. Alkhatib, K. Salloum, and S. Balloul, “Predictive analytics using big data for increased customer loyalty: Syriatel Telecom Company case study,” Journal of Big Data, vol. 7, no. 1, p. 29, Dec. 2020, doi: 10.1186/s40537-020-00290-0.
N. Yau, Visualize This: The Flowing Data Guide to Design, Visualization, and Statistics Second Edition. Indianapolis: Wiley, 2024. doi: 10.1002/9781394319817.
A. S. Tarawneh, A. B. Hassanat, G. A. Altarawneh, and A. Almuhaimeed, “Stop Oversampling for Class Imbalance Learning: A Review,” IEEE Access, vol. 10, pp. 47643–47660, 2022, doi: 10.1109/ACCESS.2022.3169512.
TensorFlow Developers, “TensorFlow,” Zenodo. Accessed: Jul. 13, 2024. [Online]. Available: https://zenodo.org/doi/10.5281/zenodo.4724125
H. Pratiwi et al., “Sigmoid Activation Function in Selecting the Best Model of Artificial Neural Networks,” Journal of Physics: Conference Series, vol. 1471, no. 1, p. 012010, Feb. 2020, doi: 10.1088/1742-6596/1471/1/012010.
C. Aliferis and G. Simon, “Overfitting, Underfitting and General Model Overconfidence and Under-Performance Pitfalls and Best Practices in Machine Learning and AI,” in Artificial Intelligence and Machine Learning in Health Care and Medical Sciences, G. J. Simon and C. Aliferis, Eds., Cham: Springer International Publishing, 2024, pp. 477–524. doi: 10.1007/978-3-031-39355-6_10.
D. J. Hand, P. Christen, and N. Kirielle, “F*: an interpretable transformation of the F-measure,” Machine Learning, vol. 110, no. 3, pp. 451–456, Mar. 2021, doi: 10.1007/s10994-021-05964-1.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2024 Mochammad Daffa Putra Karyudi, Anis Zubair
This work is licensed under a Creative Commons Attribution 4.0 International License.