Enhancing Student Performance Prediction Through Machine Learning and Deep Learning Models

  • Authors

    • Kannan M Department of Computer Science, Faculty of Science and Humanities, SRM Institute of Science and Technology, Kattankulathur Campus, Chennai, Tamilnadu, India
    • S. Albert Antony Raj Department of Computer Applications, Faculty of Science and Humanities, SRM Institute of Science and Technology, Kattankulathur, Chennai, Tamilnadu, India. alberts@srmist.edu.in
    https://doi.org/10.14419/88yz9489

    Received date: April 15, 2025

    Accepted date: June 9, 2025

    Published date: June 21, 2025

  • Student performance; Attention-based model; Kalman Filter; Convolutional layer; Recursive layer
  • Abstract

    This work addresses the challenge of predicting student performance by investigating the sparsity of student action data and the imbalance between performance and completion rates. Initially, various machine learning (ML) and deep learning (DL) methods were compared using the Swayam dataset, highlighting the superior performance of DL algorithms such as CNN and LSTM. Subsequently, an attention-based model combining convolutional and recursive layers (with LSTM units) was implemented and trained, although insufficient training data limited accurate model assessment. Lastly, the study focused on LSTM, TCN, and Kalman Filter models using the NPTEL dataset, leveraging oversampling (ADASYN) and data densification (PCA) techniques to address class imbalance and data sparsity issues. The Kalman Filter model demonstrated superior performance in terms of AUCPR, while LSTM and TCN models outperformed it in binary classification. TCN showed increased efficiency over LSTM, especially for longer time sequences. Future work will involve applying these techniques to academic datasets, potentially retraining models with ADASYN and PCA algorithms to further improve performance. Additionally, emphasis will be placed on exploring the efficacy of TCN models in solving student performance prediction tasks.

  • References

    1. Waheed, H., Hassan, S. U., Aljohani, N. R., Hardman, J., Alelyani, S., & Nawaz, R. (2020). Predicting the academic performance of students from VLE big data using deep learning models. Computers in Human Behavior, 104, 106189.
    2. Tomasevic, N., Gvozdenovic, N., & Vranes, S. (2020). An overview and comparison of supervised data mining techniques for student exam per-formance prediction. Computers & Education, 143, 103676.
    3. Wu, Z., He, T., Mao, C., & Huang, C. (2020). Exam paper generation based on the performance prediction of the student group. Information Sci-ences, 532, 72–90.
    4. Yan, J., Zhang, Z., Lin, K., Yang, F., & Luo, X. (2020). A hybrid scheme-based one-vs-all decision trees for multi-class classification tasks. Knowledge-Based Systems, 198, 105922.
    5. Khan, A., & Ghosh, S. K. (2021). Student performance analysis and prediction in classroom learning: A review of educational data mining studies. Education and Information Technologies, 26(1), 205–240.
    6. Tangirala, S. (2020). Evaluating the impact of GINI index and information gain on classification using the decision tree classifier algorithm. Inter-national Journal of Advanced Computer Science and Applications, 11(2), 612–619.
    7. Tsiakmaki, M., Kostopoulos, G., Kotsiantis, S., & Ragos, O. (2020). Implementing AutoML in educational data mining for prediction tasks. Ap-plied Sciences, 10(1), 90–117.
    8. Moises, R. G., Maria, D. P. P. R., & Francisco, O. (2020). Massive LMS log data analysis for the early prediction of course-agnostic student per-formance. Computers & Education, 163, 104083.
    9. Walsh, J. N., & Rísquez, A. (2020). Using cluster analysis to explore the engagement with a flipped classroom of native and non-native English-speaking management students. The International Journal of Management Education, 18(2), 100362.
    10. Karthikeyan, V. G., Thangaraj, P., & Karthik, S. (2020). Towards developing hybrid educational data mining model (HEDM) for effi-cient and ac-curate student performance evaluation. Soft Computing, 24(24), 18477–18487.
    11. Crivei, L. M., Czibula, G., Ciubotariu, G., & Dindelegan, M. (2020). Unsupervised learning-based mining of academic data sets for students’ per-formance analysis. Proceedings of the IEEE 14th International Symposium on Applied Computational Intelligence and In-formatics (SACI), 11–16.
    12. Delgado, S., Moran, F., Jose, J. C. S., & Burgos, D. (2021). Analysis of students’ behavior through user clustering in online learning settings based on self-organizing maps neural networks. IEEE Access, 9, 132592–132608.
    13. Okoye, K., Arrona-Palacios, A., Camacho-Zuñiga, C., Achem, J. A. G., Escamilla, J., & Hosseini, S. (2021). Towards teaching analyt-ics: A contex-tual model for analysis of students’ evaluation of teaching through text mining and machine learning classification. Education and Information Technologies, 26, 1–43.
    14. Kumar, E. S. V., Balamurugan, S. A. A., & Sasikala, S. (2021). Multi-tier student performance evaluation model (MTSPEM) with integrated classi-fication techniques for educational decision making. International Journal of Computational Intelligence Systems, 14(1), 1796–1808.
    15. Siddiqa, A., Naqvi, S. A. Z., Ahsan, M., Ditta, A., Alquhayz, H., Khan, M. A., et al. (2022). An improved evolutionary algorithm for data mining and knowledge discovery. Computer Modeling in Engineering & Sciences, 71(1), 1233–1247.
    16. Wen, Y., Tian, Y., Wen, B., Zhou, Q., Cai, G., & Liu, S. (2020). Consideration of the local correlation of learning behaviors to predict dropouts from MOOCs. Tsinghua Science and Technology, 25(3), 336–347.
    17. Lin, S. Y., Wu, C. M., Chen, S. L., & Lin, T. L. (2020). Continuous facial emotion recognition methods are based on deep learning of aca-demic emotions. Sensors and Materials, 32(10), 3243–3259.
    18. Farissi, A., Dahlan, H. M., & Samsuryadi. (2020). Genetic algorithm-based feature selection with ensemble methods for student aca-demic perfor-mance prediction. Journal of Physics: Conference Series, 1500(1), 012014.
    19. Turabieh, H., Azwari, S. A., Rokaya, M., Alosaimi, W., Alharbi, A., Alhakami, W., et al. (2021). Enhanced Harris hawk’s optimization as a feature selection for the prediction of student performance. Computing, 103, 1417–1438.
    20. Ma, H. B., Yang, S. Y., Feng, D. Z., & Jiao, L. C. (2021). Progressive mimic learning: A new perspective to train lightweight CNN model-els. Neu-rocomputing, 456, 220–231.
    21. Nabil, A., Seyam, M., & Abou-Elfetouh, A. (2021). Prediction of students’ academic performance based on courses’ grades using deep neural net-works. IEEE Access, 9, 140731–140746.
    22. Gao, L., Zhao, Z., Li, C., Zhao, J., & Zeng, Q. (2022). Deep cognitive diagnosis model for predicting students’ performance. Future Generation Computer Systems, 126, 252–262.
    23. Mishra, P., Biancolillo, A., Roger, J. M., Marini, F., & Rutledge, D. N. (2020). New data preprocessing trends based on an ensemble of multiple preprocessing techniques. TrAC Trends in Analytical Chemistry, 132, 116045.
    24. Zhao, S., Zhou, D., Wang, H., Chen, D., & Yu, L. (2025). Enhancing student academic success prediction through ensemble learning and image-based behavioral data transformation. Applied Sciences, 15(3), 1231.
    25. Hasan, R., Palaniappan, S., Mahmood, S., Abbas, A., & Sarker, K. U. (2021). Dataset of students’ performance using student infor-mation system, Moodle and the mobile application “eDify”. Data, 6(110), 1–7.
    26. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735–1780.
    27. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2023). Attention is all you need. arXiv preprint arXiv:1706.03762v7 [cs.CL]. https://doi.org/10.48550/arXiv.1706.03762
  • Downloads

  • How to Cite

    M, K., & Raj , S. A. A. . (2025). Enhancing Student Performance Prediction Through Machine Learning and Deep Learning Models. International Journal of Basic and Applied Sciences, 14(2), 296-310. https://doi.org/10.14419/88yz9489